00:00:00.001 Started by upstream project "autotest-spdk-master-vs-dpdk-v23.11" build number 1010 00:00:00.001 originally caused by: 00:00:00.001 Started by upstream project "nightly-trigger" build number 3672 00:00:00.001 originally caused by: 00:00:00.001 Started by timer 00:00:00.051 Checking out git https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool into /var/jenkins_home/workspace/raid-vg-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4 to read jbp/jenkins/jjb-config/jobs/autotest-downstream/autotest-vg.groovy 00:00:00.053 The recommended git tool is: git 00:00:00.053 using credential 00000000-0000-0000-0000-000000000002 00:00:00.056 > git rev-parse --resolve-git-dir /var/jenkins_home/workspace/raid-vg-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4/jbp/.git # timeout=10 00:00:00.067 Fetching changes from the remote Git repository 00:00:00.071 > git config remote.origin.url https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool # timeout=10 00:00:00.094 Using shallow fetch with depth 1 00:00:00.094 Fetching upstream changes from https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool 00:00:00.094 > git --version # timeout=10 00:00:00.121 > git --version # 'git version 2.39.2' 00:00:00.121 using GIT_ASKPASS to set credentials SPDKCI HTTPS Credentials 00:00:00.154 Setting http proxy: proxy-dmz.intel.com:911 00:00:00.154 > git fetch --tags --force --progress --depth=1 -- https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool refs/heads/master # timeout=5 00:00:03.426 > git rev-parse origin/FETCH_HEAD^{commit} # timeout=10 00:00:03.436 > git rev-parse FETCH_HEAD^{commit} # timeout=10 00:00:03.447 Checking out Revision db4637e8b949f278f369ec13f70585206ccd9507 (FETCH_HEAD) 00:00:03.447 > git config core.sparsecheckout # timeout=10 00:00:03.457 > git read-tree -mu HEAD # timeout=10 00:00:03.473 > git checkout -f db4637e8b949f278f369ec13f70585206ccd9507 # timeout=5 00:00:03.499 Commit message: "jenkins/jjb-config: Add missing SPDK_TEST_NVME_INTERRUPT flag" 00:00:03.499 > git rev-list --no-walk db4637e8b949f278f369ec13f70585206ccd9507 # timeout=10 00:00:03.593 [Pipeline] Start of Pipeline 00:00:03.607 [Pipeline] library 00:00:03.609 Loading library shm_lib@master 00:00:03.609 Library shm_lib@master is cached. Copying from home. 00:00:03.625 [Pipeline] node 00:00:03.650 Running on VM-host-WFP7 in /var/jenkins/workspace/raid-vg-autotest 00:00:03.652 [Pipeline] { 00:00:03.664 [Pipeline] catchError 00:00:03.666 [Pipeline] { 00:00:03.677 [Pipeline] wrap 00:00:03.685 [Pipeline] { 00:00:03.697 [Pipeline] stage 00:00:03.699 [Pipeline] { (Prologue) 00:00:03.720 [Pipeline] echo 00:00:03.722 Node: VM-host-WFP7 00:00:03.727 [Pipeline] cleanWs 00:00:03.739 [WS-CLEANUP] Deleting project workspace... 00:00:03.739 [WS-CLEANUP] Deferred wipeout is used... 00:00:03.746 [WS-CLEANUP] done 00:00:03.933 [Pipeline] setCustomBuildProperty 00:00:03.993 [Pipeline] httpRequest 00:00:04.438 [Pipeline] echo 00:00:04.441 Sorcerer 10.211.164.101 is alive 00:00:04.451 [Pipeline] retry 00:00:04.454 [Pipeline] { 00:00:04.470 [Pipeline] httpRequest 00:00:04.474 HttpMethod: GET 00:00:04.475 URL: http://10.211.164.101/packages/jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:00:04.476 Sending request to url: http://10.211.164.101/packages/jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:00:04.477 Response Code: HTTP/1.1 200 OK 00:00:04.477 Success: Status code 200 is in the accepted range: 200,404 00:00:04.478 Saving response body to /var/jenkins/workspace/raid-vg-autotest/jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:00:04.624 [Pipeline] } 00:00:04.644 [Pipeline] // retry 00:00:04.651 [Pipeline] sh 00:00:04.932 + tar --no-same-owner -xf jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:00:04.948 [Pipeline] httpRequest 00:00:07.966 [Pipeline] echo 00:00:07.968 Sorcerer 10.211.164.101 is dead 00:00:07.978 [Pipeline] httpRequest 00:00:08.493 [Pipeline] echo 00:00:08.494 Sorcerer 10.211.164.101 is alive 00:00:08.501 [Pipeline] retry 00:00:08.502 [Pipeline] { 00:00:08.510 [Pipeline] httpRequest 00:00:08.514 HttpMethod: GET 00:00:08.514 URL: http://10.211.164.101/packages/spdk_35cd3e84d4a92eacc8c9de6c2cd81450ef5bcc54.tar.gz 00:00:08.515 Sending request to url: http://10.211.164.101/packages/spdk_35cd3e84d4a92eacc8c9de6c2cd81450ef5bcc54.tar.gz 00:00:08.527 Response Code: HTTP/1.1 200 OK 00:00:08.527 Success: Status code 200 is in the accepted range: 200,404 00:00:08.527 Saving response body to /var/jenkins/workspace/raid-vg-autotest/spdk_35cd3e84d4a92eacc8c9de6c2cd81450ef5bcc54.tar.gz 00:00:54.194 [Pipeline] } 00:00:54.218 [Pipeline] // retry 00:00:54.228 [Pipeline] sh 00:00:54.520 + tar --no-same-owner -xf spdk_35cd3e84d4a92eacc8c9de6c2cd81450ef5bcc54.tar.gz 00:00:57.094 [Pipeline] sh 00:00:57.381 + git -C spdk log --oneline -n5 00:00:57.381 35cd3e84d bdev/part: Pass through dif_check_flags via dif_check_flags_exclude_mask 00:00:57.381 01a2c4855 bdev/passthru: Pass through dif_check_flags via dif_check_flags_exclude_mask 00:00:57.381 9094b9600 bdev: Assert to check if I/O pass dif_check_flags not enabled by bdev 00:00:57.381 2e10c84c8 nvmf: Expose DIF type of namespace to host again 00:00:57.381 38b931b23 nvmf: Set bdev_ext_io_opts::dif_check_flags_exclude_mask for read/write 00:00:57.402 [Pipeline] withCredentials 00:00:57.414 > git --version # timeout=10 00:00:57.428 > git --version # 'git version 2.39.2' 00:00:57.447 Masking supported pattern matches of $GIT_PASSWORD or $GIT_ASKPASS 00:00:57.449 [Pipeline] { 00:00:57.459 [Pipeline] retry 00:00:57.461 [Pipeline] { 00:00:57.478 [Pipeline] sh 00:00:57.763 + git ls-remote http://dpdk.org/git/dpdk-stable v23.11 00:00:58.036 [Pipeline] } 00:00:58.054 [Pipeline] // retry 00:00:58.060 [Pipeline] } 00:00:58.076 [Pipeline] // withCredentials 00:00:58.087 [Pipeline] httpRequest 00:00:58.538 [Pipeline] echo 00:00:58.539 Sorcerer 10.211.164.101 is alive 00:00:58.547 [Pipeline] retry 00:00:58.549 [Pipeline] { 00:00:58.560 [Pipeline] httpRequest 00:00:58.564 HttpMethod: GET 00:00:58.565 URL: http://10.211.164.101/packages/dpdk_d15625009dced269fcec27fc81dd74fd58d54cdb.tar.gz 00:00:58.565 Sending request to url: http://10.211.164.101/packages/dpdk_d15625009dced269fcec27fc81dd74fd58d54cdb.tar.gz 00:00:58.577 Response Code: HTTP/1.1 200 OK 00:00:58.577 Success: Status code 200 is in the accepted range: 200,404 00:00:58.578 Saving response body to /var/jenkins/workspace/raid-vg-autotest/dpdk_d15625009dced269fcec27fc81dd74fd58d54cdb.tar.gz 00:01:17.477 [Pipeline] } 00:01:17.493 [Pipeline] // retry 00:01:17.500 [Pipeline] sh 00:01:17.787 + tar --no-same-owner -xf dpdk_d15625009dced269fcec27fc81dd74fd58d54cdb.tar.gz 00:01:19.186 [Pipeline] sh 00:01:19.479 + git -C dpdk log --oneline -n5 00:01:19.479 eeb0605f11 version: 23.11.0 00:01:19.479 238778122a doc: update release notes for 23.11 00:01:19.479 46aa6b3cfc doc: fix description of RSS features 00:01:19.479 dd88f51a57 devtools: forbid DPDK API in cnxk base driver 00:01:19.479 7e421ae345 devtools: support skipping forbid rule check 00:01:19.517 [Pipeline] writeFile 00:01:19.536 [Pipeline] sh 00:01:19.823 + jbp/jenkins/jjb-config/jobs/scripts/autorun_quirks.sh 00:01:19.837 [Pipeline] sh 00:01:20.124 + cat autorun-spdk.conf 00:01:20.124 SPDK_RUN_FUNCTIONAL_TEST=1 00:01:20.124 SPDK_RUN_ASAN=1 00:01:20.124 SPDK_RUN_UBSAN=1 00:01:20.124 SPDK_TEST_RAID=1 00:01:20.124 SPDK_TEST_NATIVE_DPDK=v23.11 00:01:20.124 SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:01:20.124 SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:01:20.131 RUN_NIGHTLY=1 00:01:20.133 [Pipeline] } 00:01:20.148 [Pipeline] // stage 00:01:20.164 [Pipeline] stage 00:01:20.167 [Pipeline] { (Run VM) 00:01:20.180 [Pipeline] sh 00:01:20.465 + jbp/jenkins/jjb-config/jobs/scripts/prepare_nvme.sh 00:01:20.465 + echo 'Start stage prepare_nvme.sh' 00:01:20.465 Start stage prepare_nvme.sh 00:01:20.465 + [[ -n 2 ]] 00:01:20.465 + disk_prefix=ex2 00:01:20.465 + [[ -n /var/jenkins/workspace/raid-vg-autotest ]] 00:01:20.465 + [[ -e /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf ]] 00:01:20.465 + source /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf 00:01:20.465 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:01:20.465 ++ SPDK_RUN_ASAN=1 00:01:20.465 ++ SPDK_RUN_UBSAN=1 00:01:20.466 ++ SPDK_TEST_RAID=1 00:01:20.466 ++ SPDK_TEST_NATIVE_DPDK=v23.11 00:01:20.466 ++ SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:01:20.466 ++ SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:01:20.466 ++ RUN_NIGHTLY=1 00:01:20.466 + cd /var/jenkins/workspace/raid-vg-autotest 00:01:20.466 + nvme_files=() 00:01:20.466 + declare -A nvme_files 00:01:20.466 + backend_dir=/var/lib/libvirt/images/backends 00:01:20.466 + nvme_files['nvme.img']=5G 00:01:20.466 + nvme_files['nvme-cmb.img']=5G 00:01:20.466 + nvme_files['nvme-multi0.img']=4G 00:01:20.466 + nvme_files['nvme-multi1.img']=4G 00:01:20.466 + nvme_files['nvme-multi2.img']=4G 00:01:20.466 + nvme_files['nvme-openstack.img']=8G 00:01:20.466 + nvme_files['nvme-zns.img']=5G 00:01:20.466 + (( SPDK_TEST_NVME_PMR == 1 )) 00:01:20.466 + (( SPDK_TEST_FTL == 1 )) 00:01:20.466 + (( SPDK_TEST_NVME_FDP == 1 )) 00:01:20.466 + [[ ! -d /var/lib/libvirt/images/backends ]] 00:01:20.466 + for nvme in "${!nvme_files[@]}" 00:01:20.466 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex2-nvme-multi2.img -s 4G 00:01:20.466 Formatting '/var/lib/libvirt/images/backends/ex2-nvme-multi2.img', fmt=raw size=4294967296 preallocation=falloc 00:01:20.466 + for nvme in "${!nvme_files[@]}" 00:01:20.466 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex2-nvme-cmb.img -s 5G 00:01:20.466 Formatting '/var/lib/libvirt/images/backends/ex2-nvme-cmb.img', fmt=raw size=5368709120 preallocation=falloc 00:01:20.466 + for nvme in "${!nvme_files[@]}" 00:01:20.466 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex2-nvme-openstack.img -s 8G 00:01:20.466 Formatting '/var/lib/libvirt/images/backends/ex2-nvme-openstack.img', fmt=raw size=8589934592 preallocation=falloc 00:01:20.466 + for nvme in "${!nvme_files[@]}" 00:01:20.466 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex2-nvme-zns.img -s 5G 00:01:20.466 Formatting '/var/lib/libvirt/images/backends/ex2-nvme-zns.img', fmt=raw size=5368709120 preallocation=falloc 00:01:20.466 + for nvme in "${!nvme_files[@]}" 00:01:20.466 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex2-nvme-multi1.img -s 4G 00:01:20.466 Formatting '/var/lib/libvirt/images/backends/ex2-nvme-multi1.img', fmt=raw size=4294967296 preallocation=falloc 00:01:20.466 + for nvme in "${!nvme_files[@]}" 00:01:20.466 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex2-nvme-multi0.img -s 4G 00:01:20.466 Formatting '/var/lib/libvirt/images/backends/ex2-nvme-multi0.img', fmt=raw size=4294967296 preallocation=falloc 00:01:20.466 + for nvme in "${!nvme_files[@]}" 00:01:20.466 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex2-nvme.img -s 5G 00:01:21.036 Formatting '/var/lib/libvirt/images/backends/ex2-nvme.img', fmt=raw size=5368709120 preallocation=falloc 00:01:21.036 ++ sudo grep -rl ex2-nvme.img /etc/libvirt/qemu 00:01:21.036 + echo 'End stage prepare_nvme.sh' 00:01:21.036 End stage prepare_nvme.sh 00:01:21.307 [Pipeline] sh 00:01:21.590 + DISTRO=fedora39 CPUS=10 RAM=12288 jbp/jenkins/jjb-config/jobs/scripts/vagrant_create_vm.sh 00:01:21.590 Setup: -n 10 -s 12288 -x http://proxy-dmz.intel.com:911 -p libvirt --qemu-emulator=/usr/local/qemu/vanilla-v8.0.0/bin/qemu-system-x86_64 -b /var/lib/libvirt/images/backends/ex2-nvme.img -b /var/lib/libvirt/images/backends/ex2-nvme-multi0.img,nvme,/var/lib/libvirt/images/backends/ex2-nvme-multi1.img:/var/lib/libvirt/images/backends/ex2-nvme-multi2.img -H -a -v -f fedora39 00:01:21.590 00:01:21.590 DIR=/var/jenkins/workspace/raid-vg-autotest/spdk/scripts/vagrant 00:01:21.590 SPDK_DIR=/var/jenkins/workspace/raid-vg-autotest/spdk 00:01:21.590 VAGRANT_TARGET=/var/jenkins/workspace/raid-vg-autotest 00:01:21.590 HELP=0 00:01:21.590 DRY_RUN=0 00:01:21.590 NVME_FILE=/var/lib/libvirt/images/backends/ex2-nvme.img,/var/lib/libvirt/images/backends/ex2-nvme-multi0.img, 00:01:21.590 NVME_DISKS_TYPE=nvme,nvme, 00:01:21.590 NVME_AUTO_CREATE=0 00:01:21.590 NVME_DISKS_NAMESPACES=,/var/lib/libvirt/images/backends/ex2-nvme-multi1.img:/var/lib/libvirt/images/backends/ex2-nvme-multi2.img, 00:01:21.590 NVME_CMB=,, 00:01:21.590 NVME_PMR=,, 00:01:21.590 NVME_ZNS=,, 00:01:21.590 NVME_MS=,, 00:01:21.590 NVME_FDP=,, 00:01:21.590 SPDK_VAGRANT_DISTRO=fedora39 00:01:21.590 SPDK_VAGRANT_VMCPU=10 00:01:21.590 SPDK_VAGRANT_VMRAM=12288 00:01:21.590 SPDK_VAGRANT_PROVIDER=libvirt 00:01:21.590 SPDK_VAGRANT_HTTP_PROXY=http://proxy-dmz.intel.com:911 00:01:21.590 SPDK_QEMU_EMULATOR=/usr/local/qemu/vanilla-v8.0.0/bin/qemu-system-x86_64 00:01:21.590 SPDK_OPENSTACK_NETWORK=0 00:01:21.590 VAGRANT_PACKAGE_BOX=0 00:01:21.590 VAGRANTFILE=/var/jenkins/workspace/raid-vg-autotest/spdk/scripts/vagrant/Vagrantfile 00:01:21.590 FORCE_DISTRO=true 00:01:21.590 VAGRANT_BOX_VERSION= 00:01:21.590 EXTRA_VAGRANTFILES= 00:01:21.590 NIC_MODEL=virtio 00:01:21.590 00:01:21.590 mkdir: created directory '/var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt' 00:01:21.590 /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt /var/jenkins/workspace/raid-vg-autotest 00:01:23.500 Bringing machine 'default' up with 'libvirt' provider... 00:01:24.070 ==> default: Creating image (snapshot of base box volume). 00:01:24.070 ==> default: Creating domain with the following settings... 00:01:24.070 ==> default: -- Name: fedora39-39-1.5-1721788873-2326_default_1732719951_a1d83bfdfa56f53b88d0 00:01:24.070 ==> default: -- Domain type: kvm 00:01:24.070 ==> default: -- Cpus: 10 00:01:24.070 ==> default: -- Feature: acpi 00:01:24.070 ==> default: -- Feature: apic 00:01:24.070 ==> default: -- Feature: pae 00:01:24.070 ==> default: -- Memory: 12288M 00:01:24.070 ==> default: -- Memory Backing: hugepages: 00:01:24.070 ==> default: -- Management MAC: 00:01:24.070 ==> default: -- Loader: 00:01:24.070 ==> default: -- Nvram: 00:01:24.070 ==> default: -- Base box: spdk/fedora39 00:01:24.070 ==> default: -- Storage pool: default 00:01:24.070 ==> default: -- Image: /var/lib/libvirt/images/fedora39-39-1.5-1721788873-2326_default_1732719951_a1d83bfdfa56f53b88d0.img (20G) 00:01:24.070 ==> default: -- Volume Cache: default 00:01:24.070 ==> default: -- Kernel: 00:01:24.070 ==> default: -- Initrd: 00:01:24.070 ==> default: -- Graphics Type: vnc 00:01:24.070 ==> default: -- Graphics Port: -1 00:01:24.070 ==> default: -- Graphics IP: 127.0.0.1 00:01:24.070 ==> default: -- Graphics Password: Not defined 00:01:24.070 ==> default: -- Video Type: cirrus 00:01:24.070 ==> default: -- Video VRAM: 9216 00:01:24.070 ==> default: -- Sound Type: 00:01:24.070 ==> default: -- Keymap: en-us 00:01:24.070 ==> default: -- TPM Path: 00:01:24.070 ==> default: -- INPUT: type=mouse, bus=ps2 00:01:24.070 ==> default: -- Command line args: 00:01:24.070 ==> default: -> value=-device, 00:01:24.070 ==> default: -> value=nvme,id=nvme-0,serial=12340,addr=0x10, 00:01:24.070 ==> default: -> value=-drive, 00:01:24.070 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex2-nvme.img,if=none,id=nvme-0-drive0, 00:01:24.070 ==> default: -> value=-device, 00:01:24.070 ==> default: -> value=nvme-ns,drive=nvme-0-drive0,bus=nvme-0,nsid=1,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:01:24.070 ==> default: -> value=-device, 00:01:24.070 ==> default: -> value=nvme,id=nvme-1,serial=12341,addr=0x11, 00:01:24.070 ==> default: -> value=-drive, 00:01:24.070 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex2-nvme-multi0.img,if=none,id=nvme-1-drive0, 00:01:24.070 ==> default: -> value=-device, 00:01:24.070 ==> default: -> value=nvme-ns,drive=nvme-1-drive0,bus=nvme-1,nsid=1,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:01:24.070 ==> default: -> value=-drive, 00:01:24.070 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex2-nvme-multi1.img,if=none,id=nvme-1-drive1, 00:01:24.070 ==> default: -> value=-device, 00:01:24.070 ==> default: -> value=nvme-ns,drive=nvme-1-drive1,bus=nvme-1,nsid=2,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:01:24.070 ==> default: -> value=-drive, 00:01:24.070 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex2-nvme-multi2.img,if=none,id=nvme-1-drive2, 00:01:24.070 ==> default: -> value=-device, 00:01:24.070 ==> default: -> value=nvme-ns,drive=nvme-1-drive2,bus=nvme-1,nsid=3,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:01:24.331 ==> default: Creating shared folders metadata... 00:01:24.331 ==> default: Starting domain. 00:01:26.243 ==> default: Waiting for domain to get an IP address... 00:01:41.143 ==> default: Waiting for SSH to become available... 00:01:42.526 ==> default: Configuring and enabling network interfaces... 00:01:49.109 default: SSH address: 192.168.121.213:22 00:01:49.109 default: SSH username: vagrant 00:01:49.109 default: SSH auth method: private key 00:01:52.409 ==> default: Rsyncing folder: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest/spdk/ => /home/vagrant/spdk_repo/spdk 00:01:58.991 ==> default: Rsyncing folder: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest/dpdk/ => /home/vagrant/spdk_repo/dpdk 00:02:05.571 ==> default: Mounting SSHFS shared folder... 00:02:08.110 ==> default: Mounting folder via SSHFS: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest/fedora39-libvirt/output => /home/vagrant/spdk_repo/output 00:02:08.110 ==> default: Checking Mount.. 00:02:09.495 ==> default: Folder Successfully Mounted! 00:02:09.495 ==> default: Running provisioner: file... 00:02:10.431 default: ~/.gitconfig => .gitconfig 00:02:10.999 00:02:11.000 SUCCESS! 00:02:11.000 00:02:11.000 cd to /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt and type "vagrant ssh" to use. 00:02:11.000 Use vagrant "suspend" and vagrant "resume" to stop and start. 00:02:11.000 Use vagrant "destroy" followed by "rm -rf /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt" to destroy all trace of vm. 00:02:11.000 00:02:11.010 [Pipeline] } 00:02:11.024 [Pipeline] // stage 00:02:11.033 [Pipeline] dir 00:02:11.033 Running in /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt 00:02:11.035 [Pipeline] { 00:02:11.046 [Pipeline] catchError 00:02:11.048 [Pipeline] { 00:02:11.060 [Pipeline] sh 00:02:11.342 + vagrant ssh-config --host vagrant 00:02:11.342 + sed -ne /^Host/,$p 00:02:11.342 + tee ssh_conf 00:02:13.880 Host vagrant 00:02:13.880 HostName 192.168.121.213 00:02:13.880 User vagrant 00:02:13.880 Port 22 00:02:13.880 UserKnownHostsFile /dev/null 00:02:13.880 StrictHostKeyChecking no 00:02:13.880 PasswordAuthentication no 00:02:13.880 IdentityFile /var/lib/libvirt/images/.vagrant.d/boxes/spdk-VAGRANTSLASH-fedora39/39-1.5-1721788873-2326/libvirt/fedora39 00:02:13.880 IdentitiesOnly yes 00:02:13.880 LogLevel FATAL 00:02:13.880 ForwardAgent yes 00:02:13.880 ForwardX11 yes 00:02:13.880 00:02:13.895 [Pipeline] withEnv 00:02:13.898 [Pipeline] { 00:02:13.912 [Pipeline] sh 00:02:14.196 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant #!/bin/bash 00:02:14.196 source /etc/os-release 00:02:14.196 [[ -e /image.version ]] && img=$(< /image.version) 00:02:14.196 # Minimal, systemd-like check. 00:02:14.196 if [[ -e /.dockerenv ]]; then 00:02:14.196 # Clear garbage from the node's name: 00:02:14.196 # agt-er_autotest_547-896 -> autotest_547-896 00:02:14.196 # $HOSTNAME is the actual container id 00:02:14.196 agent=$HOSTNAME@${DOCKER_SWARM_PLUGIN_JENKINS_AGENT_NAME#*_} 00:02:14.196 if grep -q "/etc/hostname" /proc/self/mountinfo; then 00:02:14.196 # We can assume this is a mount from a host where container is running, 00:02:14.196 # so fetch its hostname to easily identify the target swarm worker. 00:02:14.196 container="$(< /etc/hostname) ($agent)" 00:02:14.196 else 00:02:14.196 # Fallback 00:02:14.196 container=$agent 00:02:14.196 fi 00:02:14.196 fi 00:02:14.196 echo "${NAME} ${VERSION_ID}|$(uname -r)|${img:-N/A}|${container:-N/A}" 00:02:14.196 00:02:14.468 [Pipeline] } 00:02:14.485 [Pipeline] // withEnv 00:02:14.493 [Pipeline] setCustomBuildProperty 00:02:14.507 [Pipeline] stage 00:02:14.510 [Pipeline] { (Tests) 00:02:14.527 [Pipeline] sh 00:02:14.810 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/jbp/jenkins/jjb-config/jobs/scripts/autoruner.sh vagrant@vagrant:./ 00:02:15.120 [Pipeline] sh 00:02:15.403 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/jbp/jenkins/jjb-config/jobs/scripts/pkgdep-autoruner.sh vagrant@vagrant:./ 00:02:15.679 [Pipeline] timeout 00:02:15.680 Timeout set to expire in 1 hr 30 min 00:02:15.682 [Pipeline] { 00:02:15.697 [Pipeline] sh 00:02:15.979 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant git -C spdk_repo/spdk reset --hard 00:02:16.549 HEAD is now at 35cd3e84d bdev/part: Pass through dif_check_flags via dif_check_flags_exclude_mask 00:02:16.561 [Pipeline] sh 00:02:16.844 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant sudo chown vagrant:vagrant spdk_repo 00:02:17.118 [Pipeline] sh 00:02:17.401 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf vagrant@vagrant:spdk_repo 00:02:17.679 [Pipeline] sh 00:02:17.963 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant JOB_BASE_NAME=raid-vg-autotest ./autoruner.sh spdk_repo 00:02:18.222 ++ readlink -f spdk_repo 00:02:18.222 + DIR_ROOT=/home/vagrant/spdk_repo 00:02:18.222 + [[ -n /home/vagrant/spdk_repo ]] 00:02:18.222 + DIR_SPDK=/home/vagrant/spdk_repo/spdk 00:02:18.222 + DIR_OUTPUT=/home/vagrant/spdk_repo/output 00:02:18.222 + [[ -d /home/vagrant/spdk_repo/spdk ]] 00:02:18.222 + [[ ! -d /home/vagrant/spdk_repo/output ]] 00:02:18.222 + [[ -d /home/vagrant/spdk_repo/output ]] 00:02:18.222 + [[ raid-vg-autotest == pkgdep-* ]] 00:02:18.222 + cd /home/vagrant/spdk_repo 00:02:18.222 + source /etc/os-release 00:02:18.222 ++ NAME='Fedora Linux' 00:02:18.222 ++ VERSION='39 (Cloud Edition)' 00:02:18.222 ++ ID=fedora 00:02:18.222 ++ VERSION_ID=39 00:02:18.222 ++ VERSION_CODENAME= 00:02:18.222 ++ PLATFORM_ID=platform:f39 00:02:18.222 ++ PRETTY_NAME='Fedora Linux 39 (Cloud Edition)' 00:02:18.222 ++ ANSI_COLOR='0;38;2;60;110;180' 00:02:18.222 ++ LOGO=fedora-logo-icon 00:02:18.222 ++ CPE_NAME=cpe:/o:fedoraproject:fedora:39 00:02:18.222 ++ HOME_URL=https://fedoraproject.org/ 00:02:18.222 ++ DOCUMENTATION_URL=https://docs.fedoraproject.org/en-US/fedora/f39/system-administrators-guide/ 00:02:18.222 ++ SUPPORT_URL=https://ask.fedoraproject.org/ 00:02:18.222 ++ BUG_REPORT_URL=https://bugzilla.redhat.com/ 00:02:18.222 ++ REDHAT_BUGZILLA_PRODUCT=Fedora 00:02:18.222 ++ REDHAT_BUGZILLA_PRODUCT_VERSION=39 00:02:18.222 ++ REDHAT_SUPPORT_PRODUCT=Fedora 00:02:18.222 ++ REDHAT_SUPPORT_PRODUCT_VERSION=39 00:02:18.222 ++ SUPPORT_END=2024-11-12 00:02:18.222 ++ VARIANT='Cloud Edition' 00:02:18.222 ++ VARIANT_ID=cloud 00:02:18.222 + uname -a 00:02:18.222 Linux fedora39-cloud-1721788873-2326 6.8.9-200.fc39.x86_64 #1 SMP PREEMPT_DYNAMIC Wed Jul 24 03:04:40 UTC 2024 x86_64 GNU/Linux 00:02:18.222 + sudo /home/vagrant/spdk_repo/spdk/scripts/setup.sh status 00:02:18.792 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:02:18.792 Hugepages 00:02:18.792 node hugesize free / total 00:02:18.792 node0 1048576kB 0 / 0 00:02:18.792 node0 2048kB 0 / 0 00:02:18.792 00:02:18.792 Type BDF Vendor Device NUMA Driver Device Block devices 00:02:18.792 virtio 0000:00:03.0 1af4 1001 unknown virtio-pci - vda 00:02:18.792 NVMe 0000:00:10.0 1b36 0010 unknown nvme nvme0 nvme0n1 00:02:18.792 NVMe 0000:00:11.0 1b36 0010 unknown nvme nvme1 nvme1n1 nvme1n2 nvme1n3 00:02:18.792 + rm -f /tmp/spdk-ld-path 00:02:18.792 + source autorun-spdk.conf 00:02:18.792 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:02:18.792 ++ SPDK_RUN_ASAN=1 00:02:18.792 ++ SPDK_RUN_UBSAN=1 00:02:18.792 ++ SPDK_TEST_RAID=1 00:02:18.792 ++ SPDK_TEST_NATIVE_DPDK=v23.11 00:02:18.792 ++ SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:02:18.792 ++ SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:02:18.792 ++ RUN_NIGHTLY=1 00:02:18.792 + (( SPDK_TEST_NVME_CMB == 1 || SPDK_TEST_NVME_PMR == 1 )) 00:02:18.792 + [[ -n '' ]] 00:02:18.792 + sudo git config --global --add safe.directory /home/vagrant/spdk_repo/spdk 00:02:18.792 + for M in /var/spdk/build-*-manifest.txt 00:02:18.792 + [[ -f /var/spdk/build-kernel-manifest.txt ]] 00:02:18.792 + cp /var/spdk/build-kernel-manifest.txt /home/vagrant/spdk_repo/output/ 00:02:18.792 + for M in /var/spdk/build-*-manifest.txt 00:02:18.792 + [[ -f /var/spdk/build-pkg-manifest.txt ]] 00:02:18.792 + cp /var/spdk/build-pkg-manifest.txt /home/vagrant/spdk_repo/output/ 00:02:18.792 + for M in /var/spdk/build-*-manifest.txt 00:02:18.792 + [[ -f /var/spdk/build-repo-manifest.txt ]] 00:02:18.792 + cp /var/spdk/build-repo-manifest.txt /home/vagrant/spdk_repo/output/ 00:02:19.051 ++ uname 00:02:19.051 + [[ Linux == \L\i\n\u\x ]] 00:02:19.051 + sudo dmesg -T 00:02:19.051 + sudo dmesg --clear 00:02:19.051 + dmesg_pid=6172 00:02:19.051 + [[ Fedora Linux == FreeBSD ]] 00:02:19.051 + sudo dmesg -Tw 00:02:19.051 + export UNBIND_ENTIRE_IOMMU_GROUP=yes 00:02:19.051 + UNBIND_ENTIRE_IOMMU_GROUP=yes 00:02:19.051 + [[ -e /var/spdk/dependencies/vhost/spdk_test_image.qcow2 ]] 00:02:19.051 + [[ -x /usr/src/fio-static/fio ]] 00:02:19.051 + export FIO_BIN=/usr/src/fio-static/fio 00:02:19.051 + FIO_BIN=/usr/src/fio-static/fio 00:02:19.051 + [[ '' == \/\q\e\m\u\_\v\f\i\o\/* ]] 00:02:19.051 + [[ ! -v VFIO_QEMU_BIN ]] 00:02:19.051 + [[ -e /usr/local/qemu/vfio-user-latest ]] 00:02:19.051 + export VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:02:19.051 + VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:02:19.051 + [[ -e /usr/local/qemu/vanilla-latest ]] 00:02:19.051 + export QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:02:19.051 + QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:02:19.051 + spdk/autorun.sh /home/vagrant/spdk_repo/autorun-spdk.conf 00:02:19.051 15:06:47 -- common/autotest_common.sh@1692 -- $ [[ n == y ]] 00:02:19.051 15:06:47 -- spdk/autorun.sh@20 -- $ source /home/vagrant/spdk_repo/autorun-spdk.conf 00:02:19.051 15:06:47 -- spdk_repo/autorun-spdk.conf@1 -- $ SPDK_RUN_FUNCTIONAL_TEST=1 00:02:19.051 15:06:47 -- spdk_repo/autorun-spdk.conf@2 -- $ SPDK_RUN_ASAN=1 00:02:19.051 15:06:47 -- spdk_repo/autorun-spdk.conf@3 -- $ SPDK_RUN_UBSAN=1 00:02:19.051 15:06:47 -- spdk_repo/autorun-spdk.conf@4 -- $ SPDK_TEST_RAID=1 00:02:19.051 15:06:47 -- spdk_repo/autorun-spdk.conf@5 -- $ SPDK_TEST_NATIVE_DPDK=v23.11 00:02:19.051 15:06:47 -- spdk_repo/autorun-spdk.conf@6 -- $ SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:02:19.051 15:06:47 -- spdk_repo/autorun-spdk.conf@7 -- $ SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:02:19.051 15:06:47 -- spdk_repo/autorun-spdk.conf@8 -- $ RUN_NIGHTLY=1 00:02:19.051 15:06:47 -- spdk/autorun.sh@22 -- $ trap 'timing_finish || exit 1' EXIT 00:02:19.051 15:06:47 -- spdk/autorun.sh@25 -- $ /home/vagrant/spdk_repo/spdk/autobuild.sh /home/vagrant/spdk_repo/autorun-spdk.conf 00:02:19.309 15:06:47 -- common/autotest_common.sh@1692 -- $ [[ n == y ]] 00:02:19.309 15:06:47 -- common/autobuild_common.sh@15 -- $ source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:02:19.309 15:06:47 -- scripts/common.sh@15 -- $ shopt -s extglob 00:02:19.309 15:06:47 -- scripts/common.sh@544 -- $ [[ -e /bin/wpdk_common.sh ]] 00:02:19.309 15:06:47 -- scripts/common.sh@552 -- $ [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:02:19.309 15:06:47 -- scripts/common.sh@553 -- $ source /etc/opt/spdk-pkgdep/paths/export.sh 00:02:19.310 15:06:47 -- paths/export.sh@2 -- $ PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:19.310 15:06:47 -- paths/export.sh@3 -- $ PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:19.310 15:06:47 -- paths/export.sh@4 -- $ PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:19.310 15:06:47 -- paths/export.sh@5 -- $ export PATH 00:02:19.310 15:06:47 -- paths/export.sh@6 -- $ echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:19.310 15:06:47 -- common/autobuild_common.sh@492 -- $ out=/home/vagrant/spdk_repo/spdk/../output 00:02:19.310 15:06:47 -- common/autobuild_common.sh@493 -- $ date +%s 00:02:19.310 15:06:47 -- common/autobuild_common.sh@493 -- $ mktemp -dt spdk_1732720007.XXXXXX 00:02:19.310 15:06:47 -- common/autobuild_common.sh@493 -- $ SPDK_WORKSPACE=/tmp/spdk_1732720007.C69Suj 00:02:19.310 15:06:47 -- common/autobuild_common.sh@495 -- $ [[ -n '' ]] 00:02:19.310 15:06:47 -- common/autobuild_common.sh@499 -- $ '[' -n v23.11 ']' 00:02:19.310 15:06:47 -- common/autobuild_common.sh@500 -- $ dirname /home/vagrant/spdk_repo/dpdk/build 00:02:19.310 15:06:47 -- common/autobuild_common.sh@500 -- $ scanbuild_exclude=' --exclude /home/vagrant/spdk_repo/dpdk' 00:02:19.310 15:06:47 -- common/autobuild_common.sh@506 -- $ scanbuild_exclude+=' --exclude /home/vagrant/spdk_repo/spdk/xnvme --exclude /tmp' 00:02:19.310 15:06:47 -- common/autobuild_common.sh@508 -- $ scanbuild='scan-build -o /home/vagrant/spdk_repo/spdk/../output/scan-build-tmp --exclude /home/vagrant/spdk_repo/dpdk --exclude /home/vagrant/spdk_repo/spdk/xnvme --exclude /tmp --status-bugs' 00:02:19.310 15:06:47 -- common/autobuild_common.sh@509 -- $ get_config_params 00:02:19.310 15:06:47 -- common/autotest_common.sh@409 -- $ xtrace_disable 00:02:19.310 15:06:47 -- common/autotest_common.sh@10 -- $ set +x 00:02:19.310 15:06:47 -- common/autobuild_common.sh@509 -- $ config_params='--enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-asan --enable-coverage --with-ublk --with-raid5f --with-dpdk=/home/vagrant/spdk_repo/dpdk/build' 00:02:19.310 15:06:47 -- common/autobuild_common.sh@511 -- $ start_monitor_resources 00:02:19.310 15:06:47 -- pm/common@17 -- $ local monitor 00:02:19.310 15:06:47 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:02:19.310 15:06:47 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:02:19.310 15:06:47 -- pm/common@25 -- $ sleep 1 00:02:19.310 15:06:47 -- pm/common@21 -- $ date +%s 00:02:19.310 15:06:47 -- pm/common@21 -- $ date +%s 00:02:19.310 15:06:47 -- pm/common@21 -- $ /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-cpu-load -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autobuild.sh.1732720007 00:02:19.310 15:06:47 -- pm/common@21 -- $ /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-vmstat -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autobuild.sh.1732720007 00:02:19.310 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autobuild.sh.1732720007_collect-vmstat.pm.log 00:02:19.310 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autobuild.sh.1732720007_collect-cpu-load.pm.log 00:02:20.250 15:06:48 -- common/autobuild_common.sh@512 -- $ trap stop_monitor_resources EXIT 00:02:20.250 15:06:48 -- spdk/autobuild.sh@11 -- $ SPDK_TEST_AUTOBUILD= 00:02:20.250 15:06:48 -- spdk/autobuild.sh@12 -- $ umask 022 00:02:20.250 15:06:48 -- spdk/autobuild.sh@13 -- $ cd /home/vagrant/spdk_repo/spdk 00:02:20.250 15:06:48 -- spdk/autobuild.sh@16 -- $ date -u 00:02:20.250 Wed Nov 27 03:06:48 PM UTC 2024 00:02:20.250 15:06:48 -- spdk/autobuild.sh@17 -- $ git describe --tags 00:02:20.250 v25.01-pre-276-g35cd3e84d 00:02:20.250 15:06:48 -- spdk/autobuild.sh@19 -- $ '[' 1 -eq 1 ']' 00:02:20.250 15:06:48 -- spdk/autobuild.sh@20 -- $ run_test asan echo 'using asan' 00:02:20.250 15:06:48 -- common/autotest_common.sh@1105 -- $ '[' 3 -le 1 ']' 00:02:20.250 15:06:48 -- common/autotest_common.sh@1111 -- $ xtrace_disable 00:02:20.250 15:06:48 -- common/autotest_common.sh@10 -- $ set +x 00:02:20.250 ************************************ 00:02:20.250 START TEST asan 00:02:20.250 ************************************ 00:02:20.250 using asan 00:02:20.250 15:06:48 asan -- common/autotest_common.sh@1129 -- $ echo 'using asan' 00:02:20.250 00:02:20.250 real 0m0.001s 00:02:20.250 user 0m0.000s 00:02:20.250 sys 0m0.000s 00:02:20.250 15:06:48 asan -- common/autotest_common.sh@1130 -- $ xtrace_disable 00:02:20.250 15:06:48 asan -- common/autotest_common.sh@10 -- $ set +x 00:02:20.250 ************************************ 00:02:20.250 END TEST asan 00:02:20.250 ************************************ 00:02:20.250 15:06:48 -- spdk/autobuild.sh@23 -- $ '[' 1 -eq 1 ']' 00:02:20.250 15:06:48 -- spdk/autobuild.sh@24 -- $ run_test ubsan echo 'using ubsan' 00:02:20.250 15:06:48 -- common/autotest_common.sh@1105 -- $ '[' 3 -le 1 ']' 00:02:20.250 15:06:48 -- common/autotest_common.sh@1111 -- $ xtrace_disable 00:02:20.250 15:06:48 -- common/autotest_common.sh@10 -- $ set +x 00:02:20.250 ************************************ 00:02:20.250 START TEST ubsan 00:02:20.250 ************************************ 00:02:20.250 using ubsan 00:02:20.250 15:06:48 ubsan -- common/autotest_common.sh@1129 -- $ echo 'using ubsan' 00:02:20.250 00:02:20.250 real 0m0.000s 00:02:20.250 user 0m0.000s 00:02:20.250 sys 0m0.000s 00:02:20.250 15:06:48 ubsan -- common/autotest_common.sh@1130 -- $ xtrace_disable 00:02:20.250 15:06:48 ubsan -- common/autotest_common.sh@10 -- $ set +x 00:02:20.250 ************************************ 00:02:20.250 END TEST ubsan 00:02:20.250 ************************************ 00:02:20.511 15:06:48 -- spdk/autobuild.sh@27 -- $ '[' -n v23.11 ']' 00:02:20.511 15:06:48 -- spdk/autobuild.sh@28 -- $ build_native_dpdk 00:02:20.511 15:06:48 -- common/autobuild_common.sh@449 -- $ run_test build_native_dpdk _build_native_dpdk 00:02:20.511 15:06:48 -- common/autotest_common.sh@1105 -- $ '[' 2 -le 1 ']' 00:02:20.511 15:06:48 -- common/autotest_common.sh@1111 -- $ xtrace_disable 00:02:20.511 15:06:48 -- common/autotest_common.sh@10 -- $ set +x 00:02:20.511 ************************************ 00:02:20.511 START TEST build_native_dpdk 00:02:20.511 ************************************ 00:02:20.511 15:06:48 build_native_dpdk -- common/autotest_common.sh@1129 -- $ _build_native_dpdk 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@48 -- $ local external_dpdk_dir 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@49 -- $ local external_dpdk_base_dir 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@50 -- $ local compiler_version 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@51 -- $ local compiler 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@52 -- $ local dpdk_kmods 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@53 -- $ local repo=dpdk 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@55 -- $ compiler=gcc 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@61 -- $ export CC=gcc 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@61 -- $ CC=gcc 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@63 -- $ [[ gcc != *clang* ]] 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@63 -- $ [[ gcc != *gcc* ]] 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@68 -- $ gcc -dumpversion 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@68 -- $ compiler_version=13 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@69 -- $ compiler_version=13 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@70 -- $ external_dpdk_dir=/home/vagrant/spdk_repo/dpdk/build 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@71 -- $ dirname /home/vagrant/spdk_repo/dpdk/build 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@71 -- $ external_dpdk_base_dir=/home/vagrant/spdk_repo/dpdk 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@73 -- $ [[ ! -d /home/vagrant/spdk_repo/dpdk ]] 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@82 -- $ orgdir=/home/vagrant/spdk_repo/spdk 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@83 -- $ git -C /home/vagrant/spdk_repo/dpdk log --oneline -n 5 00:02:20.511 eeb0605f11 version: 23.11.0 00:02:20.511 238778122a doc: update release notes for 23.11 00:02:20.511 46aa6b3cfc doc: fix description of RSS features 00:02:20.511 dd88f51a57 devtools: forbid DPDK API in cnxk base driver 00:02:20.511 7e421ae345 devtools: support skipping forbid rule check 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@85 -- $ dpdk_cflags='-fPIC -g -fcommon' 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@86 -- $ dpdk_ldflags= 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@87 -- $ dpdk_ver=23.11.0 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@89 -- $ [[ gcc == *gcc* ]] 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@89 -- $ [[ 13 -ge 5 ]] 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@90 -- $ dpdk_cflags+=' -Werror' 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@93 -- $ [[ gcc == *gcc* ]] 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@93 -- $ [[ 13 -ge 10 ]] 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@94 -- $ dpdk_cflags+=' -Wno-stringop-overflow' 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@102 -- $ DPDK_DRIVERS=("bus" "bus/pci" "bus/vdev" "mempool/ring" "net/i40e" "net/i40e/base" "power/acpi" "power/amd_pstate" "power/cppc" "power/intel_pstate" "power/intel_uncore" "power/kvm_vm") 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@103 -- $ local mlx5_libs_added=n 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@104 -- $ [[ 0 -eq 1 ]] 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@104 -- $ [[ 0 -eq 1 ]] 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@146 -- $ [[ 0 -eq 1 ]] 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@174 -- $ cd /home/vagrant/spdk_repo/dpdk 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@175 -- $ uname -s 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@175 -- $ '[' Linux = Linux ']' 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@176 -- $ lt 23.11.0 21.11.0 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@373 -- $ cmp_versions 23.11.0 '<' 21.11.0 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@336 -- $ IFS=.-: 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@336 -- $ read -ra ver1 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@337 -- $ IFS=.-: 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@337 -- $ read -ra ver2 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@338 -- $ local 'op=<' 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@340 -- $ ver1_l=3 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@341 -- $ ver2_l=3 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@344 -- $ case "$op" in 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@345 -- $ : 1 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@364 -- $ (( v = 0 )) 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@365 -- $ decimal 23 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@353 -- $ local d=23 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 23 =~ ^[0-9]+$ ]] 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@355 -- $ echo 23 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@365 -- $ ver1[v]=23 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@366 -- $ decimal 21 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@353 -- $ local d=21 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 21 =~ ^[0-9]+$ ]] 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@355 -- $ echo 21 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@366 -- $ ver2[v]=21 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@367 -- $ return 1 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@180 -- $ patch -p1 00:02:20.511 patching file config/rte_config.h 00:02:20.511 Hunk #1 succeeded at 60 (offset 1 line). 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@183 -- $ lt 23.11.0 24.07.0 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@373 -- $ cmp_versions 23.11.0 '<' 24.07.0 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@336 -- $ IFS=.-: 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@336 -- $ read -ra ver1 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@337 -- $ IFS=.-: 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@337 -- $ read -ra ver2 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@338 -- $ local 'op=<' 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@340 -- $ ver1_l=3 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@341 -- $ ver2_l=3 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@344 -- $ case "$op" in 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@345 -- $ : 1 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@364 -- $ (( v = 0 )) 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@365 -- $ decimal 23 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@353 -- $ local d=23 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 23 =~ ^[0-9]+$ ]] 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@355 -- $ echo 23 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@365 -- $ ver1[v]=23 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@366 -- $ decimal 24 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@353 -- $ local d=24 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 24 =~ ^[0-9]+$ ]] 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@355 -- $ echo 24 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@366 -- $ ver2[v]=24 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@368 -- $ (( ver1[v] < ver2[v] )) 00:02:20.511 15:06:48 build_native_dpdk -- scripts/common.sh@368 -- $ return 0 00:02:20.511 15:06:48 build_native_dpdk -- common/autobuild_common.sh@184 -- $ patch -p1 00:02:20.511 patching file lib/pcapng/rte_pcapng.c 00:02:20.512 15:06:48 build_native_dpdk -- common/autobuild_common.sh@186 -- $ ge 23.11.0 24.07.0 00:02:20.512 15:06:48 build_native_dpdk -- scripts/common.sh@376 -- $ cmp_versions 23.11.0 '>=' 24.07.0 00:02:20.512 15:06:48 build_native_dpdk -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:02:20.512 15:06:48 build_native_dpdk -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:02:20.512 15:06:48 build_native_dpdk -- scripts/common.sh@336 -- $ IFS=.-: 00:02:20.512 15:06:48 build_native_dpdk -- scripts/common.sh@336 -- $ read -ra ver1 00:02:20.512 15:06:48 build_native_dpdk -- scripts/common.sh@337 -- $ IFS=.-: 00:02:20.512 15:06:48 build_native_dpdk -- scripts/common.sh@337 -- $ read -ra ver2 00:02:20.512 15:06:48 build_native_dpdk -- scripts/common.sh@338 -- $ local 'op=>=' 00:02:20.512 15:06:48 build_native_dpdk -- scripts/common.sh@340 -- $ ver1_l=3 00:02:20.512 15:06:48 build_native_dpdk -- scripts/common.sh@341 -- $ ver2_l=3 00:02:20.512 15:06:48 build_native_dpdk -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:02:20.512 15:06:48 build_native_dpdk -- scripts/common.sh@344 -- $ case "$op" in 00:02:20.512 15:06:48 build_native_dpdk -- scripts/common.sh@348 -- $ : 1 00:02:20.512 15:06:48 build_native_dpdk -- scripts/common.sh@364 -- $ (( v = 0 )) 00:02:20.512 15:06:48 build_native_dpdk -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:02:20.512 15:06:48 build_native_dpdk -- scripts/common.sh@365 -- $ decimal 23 00:02:20.512 15:06:48 build_native_dpdk -- scripts/common.sh@353 -- $ local d=23 00:02:20.512 15:06:48 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 23 =~ ^[0-9]+$ ]] 00:02:20.512 15:06:48 build_native_dpdk -- scripts/common.sh@355 -- $ echo 23 00:02:20.512 15:06:48 build_native_dpdk -- scripts/common.sh@365 -- $ ver1[v]=23 00:02:20.512 15:06:48 build_native_dpdk -- scripts/common.sh@366 -- $ decimal 24 00:02:20.512 15:06:48 build_native_dpdk -- scripts/common.sh@353 -- $ local d=24 00:02:20.512 15:06:48 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 24 =~ ^[0-9]+$ ]] 00:02:20.512 15:06:48 build_native_dpdk -- scripts/common.sh@355 -- $ echo 24 00:02:20.512 15:06:48 build_native_dpdk -- scripts/common.sh@366 -- $ ver2[v]=24 00:02:20.512 15:06:48 build_native_dpdk -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:02:20.512 15:06:48 build_native_dpdk -- scripts/common.sh@368 -- $ (( ver1[v] < ver2[v] )) 00:02:20.512 15:06:48 build_native_dpdk -- scripts/common.sh@368 -- $ return 1 00:02:20.512 15:06:48 build_native_dpdk -- common/autobuild_common.sh@190 -- $ dpdk_kmods=false 00:02:20.512 15:06:48 build_native_dpdk -- common/autobuild_common.sh@191 -- $ uname -s 00:02:20.512 15:06:48 build_native_dpdk -- common/autobuild_common.sh@191 -- $ '[' Linux = FreeBSD ']' 00:02:20.512 15:06:48 build_native_dpdk -- common/autobuild_common.sh@195 -- $ printf %s, bus bus/pci bus/vdev mempool/ring net/i40e net/i40e/base power/acpi power/amd_pstate power/cppc power/intel_pstate power/intel_uncore power/kvm_vm 00:02:20.512 15:06:48 build_native_dpdk -- common/autobuild_common.sh@195 -- $ meson build-tmp --prefix=/home/vagrant/spdk_repo/dpdk/build --libdir lib -Denable_docs=false -Denable_kmods=false -Dtests=false -Dc_link_args= '-Dc_args=-fPIC -g -fcommon -Werror -Wno-stringop-overflow' -Dmachine=native -Denable_drivers=bus,bus/pci,bus/vdev,mempool/ring,net/i40e,net/i40e/base,power/acpi,power/amd_pstate,power/cppc,power/intel_pstate,power/intel_uncore,power/kvm_vm, 00:02:25.797 The Meson build system 00:02:25.797 Version: 1.5.0 00:02:25.797 Source dir: /home/vagrant/spdk_repo/dpdk 00:02:25.797 Build dir: /home/vagrant/spdk_repo/dpdk/build-tmp 00:02:25.797 Build type: native build 00:02:25.797 Program cat found: YES (/usr/bin/cat) 00:02:25.797 Project name: DPDK 00:02:25.797 Project version: 23.11.0 00:02:25.797 C compiler for the host machine: gcc (gcc 13.3.1 "gcc (GCC) 13.3.1 20240522 (Red Hat 13.3.1-1)") 00:02:25.797 C linker for the host machine: gcc ld.bfd 2.40-14 00:02:25.797 Host machine cpu family: x86_64 00:02:25.797 Host machine cpu: x86_64 00:02:25.797 Message: ## Building in Developer Mode ## 00:02:25.797 Program pkg-config found: YES (/usr/bin/pkg-config) 00:02:25.797 Program check-symbols.sh found: YES (/home/vagrant/spdk_repo/dpdk/buildtools/check-symbols.sh) 00:02:25.797 Program options-ibverbs-static.sh found: YES (/home/vagrant/spdk_repo/dpdk/buildtools/options-ibverbs-static.sh) 00:02:25.797 Program python3 found: YES (/usr/bin/python3) 00:02:25.797 Program cat found: YES (/usr/bin/cat) 00:02:25.797 config/meson.build:113: WARNING: The "machine" option is deprecated. Please use "cpu_instruction_set" instead. 00:02:25.797 Compiler for C supports arguments -march=native: YES 00:02:25.797 Checking for size of "void *" : 8 00:02:25.797 Checking for size of "void *" : 8 (cached) 00:02:25.797 Library m found: YES 00:02:25.797 Library numa found: YES 00:02:25.797 Has header "numaif.h" : YES 00:02:25.797 Library fdt found: NO 00:02:25.797 Library execinfo found: NO 00:02:25.797 Has header "execinfo.h" : YES 00:02:25.797 Found pkg-config: YES (/usr/bin/pkg-config) 1.9.5 00:02:25.797 Run-time dependency libarchive found: NO (tried pkgconfig) 00:02:25.797 Run-time dependency libbsd found: NO (tried pkgconfig) 00:02:25.797 Run-time dependency jansson found: NO (tried pkgconfig) 00:02:25.797 Run-time dependency openssl found: YES 3.1.1 00:02:25.797 Run-time dependency libpcap found: YES 1.10.4 00:02:25.797 Has header "pcap.h" with dependency libpcap: YES 00:02:25.797 Compiler for C supports arguments -Wcast-qual: YES 00:02:25.797 Compiler for C supports arguments -Wdeprecated: YES 00:02:25.797 Compiler for C supports arguments -Wformat: YES 00:02:25.797 Compiler for C supports arguments -Wformat-nonliteral: NO 00:02:25.797 Compiler for C supports arguments -Wformat-security: NO 00:02:25.797 Compiler for C supports arguments -Wmissing-declarations: YES 00:02:25.797 Compiler for C supports arguments -Wmissing-prototypes: YES 00:02:25.797 Compiler for C supports arguments -Wnested-externs: YES 00:02:25.797 Compiler for C supports arguments -Wold-style-definition: YES 00:02:25.797 Compiler for C supports arguments -Wpointer-arith: YES 00:02:25.797 Compiler for C supports arguments -Wsign-compare: YES 00:02:25.797 Compiler for C supports arguments -Wstrict-prototypes: YES 00:02:25.797 Compiler for C supports arguments -Wundef: YES 00:02:25.797 Compiler for C supports arguments -Wwrite-strings: YES 00:02:25.797 Compiler for C supports arguments -Wno-address-of-packed-member: YES 00:02:25.797 Compiler for C supports arguments -Wno-packed-not-aligned: YES 00:02:25.797 Compiler for C supports arguments -Wno-missing-field-initializers: YES 00:02:25.797 Compiler for C supports arguments -Wno-zero-length-bounds: YES 00:02:25.797 Program objdump found: YES (/usr/bin/objdump) 00:02:25.797 Compiler for C supports arguments -mavx512f: YES 00:02:25.797 Checking if "AVX512 checking" compiles: YES 00:02:25.797 Fetching value of define "__SSE4_2__" : 1 00:02:25.797 Fetching value of define "__AES__" : 1 00:02:25.797 Fetching value of define "__AVX__" : 1 00:02:25.797 Fetching value of define "__AVX2__" : 1 00:02:25.797 Fetching value of define "__AVX512BW__" : 1 00:02:25.797 Fetching value of define "__AVX512CD__" : 1 00:02:25.797 Fetching value of define "__AVX512DQ__" : 1 00:02:25.797 Fetching value of define "__AVX512F__" : 1 00:02:25.797 Fetching value of define "__AVX512VL__" : 1 00:02:25.797 Fetching value of define "__PCLMUL__" : 1 00:02:25.797 Fetching value of define "__RDRND__" : 1 00:02:25.797 Fetching value of define "__RDSEED__" : 1 00:02:25.797 Fetching value of define "__VPCLMULQDQ__" : (undefined) 00:02:25.797 Fetching value of define "__znver1__" : (undefined) 00:02:25.797 Fetching value of define "__znver2__" : (undefined) 00:02:25.797 Fetching value of define "__znver3__" : (undefined) 00:02:25.797 Fetching value of define "__znver4__" : (undefined) 00:02:25.797 Compiler for C supports arguments -Wno-format-truncation: YES 00:02:25.797 Message: lib/log: Defining dependency "log" 00:02:25.797 Message: lib/kvargs: Defining dependency "kvargs" 00:02:25.797 Message: lib/telemetry: Defining dependency "telemetry" 00:02:25.797 Checking for function "getentropy" : NO 00:02:25.797 Message: lib/eal: Defining dependency "eal" 00:02:25.797 Message: lib/ring: Defining dependency "ring" 00:02:25.797 Message: lib/rcu: Defining dependency "rcu" 00:02:25.798 Message: lib/mempool: Defining dependency "mempool" 00:02:25.798 Message: lib/mbuf: Defining dependency "mbuf" 00:02:25.798 Fetching value of define "__PCLMUL__" : 1 (cached) 00:02:25.798 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:25.798 Fetching value of define "__AVX512BW__" : 1 (cached) 00:02:25.798 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:02:25.798 Fetching value of define "__AVX512VL__" : 1 (cached) 00:02:25.798 Fetching value of define "__VPCLMULQDQ__" : (undefined) (cached) 00:02:25.798 Compiler for C supports arguments -mpclmul: YES 00:02:25.798 Compiler for C supports arguments -maes: YES 00:02:25.798 Compiler for C supports arguments -mavx512f: YES (cached) 00:02:25.798 Compiler for C supports arguments -mavx512bw: YES 00:02:25.798 Compiler for C supports arguments -mavx512dq: YES 00:02:25.798 Compiler for C supports arguments -mavx512vl: YES 00:02:25.798 Compiler for C supports arguments -mvpclmulqdq: YES 00:02:25.798 Compiler for C supports arguments -mavx2: YES 00:02:25.798 Compiler for C supports arguments -mavx: YES 00:02:25.798 Message: lib/net: Defining dependency "net" 00:02:25.798 Message: lib/meter: Defining dependency "meter" 00:02:25.798 Message: lib/ethdev: Defining dependency "ethdev" 00:02:25.798 Message: lib/pci: Defining dependency "pci" 00:02:25.798 Message: lib/cmdline: Defining dependency "cmdline" 00:02:25.798 Message: lib/metrics: Defining dependency "metrics" 00:02:25.798 Message: lib/hash: Defining dependency "hash" 00:02:25.798 Message: lib/timer: Defining dependency "timer" 00:02:25.798 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:25.798 Fetching value of define "__AVX512VL__" : 1 (cached) 00:02:25.798 Fetching value of define "__AVX512CD__" : 1 (cached) 00:02:25.798 Fetching value of define "__AVX512BW__" : 1 (cached) 00:02:25.798 Message: lib/acl: Defining dependency "acl" 00:02:25.798 Message: lib/bbdev: Defining dependency "bbdev" 00:02:25.798 Message: lib/bitratestats: Defining dependency "bitratestats" 00:02:25.798 Run-time dependency libelf found: YES 0.191 00:02:25.798 Message: lib/bpf: Defining dependency "bpf" 00:02:25.798 Message: lib/cfgfile: Defining dependency "cfgfile" 00:02:25.798 Message: lib/compressdev: Defining dependency "compressdev" 00:02:25.798 Message: lib/cryptodev: Defining dependency "cryptodev" 00:02:25.798 Message: lib/distributor: Defining dependency "distributor" 00:02:25.798 Message: lib/dmadev: Defining dependency "dmadev" 00:02:25.798 Message: lib/efd: Defining dependency "efd" 00:02:25.798 Message: lib/eventdev: Defining dependency "eventdev" 00:02:25.798 Message: lib/dispatcher: Defining dependency "dispatcher" 00:02:25.798 Message: lib/gpudev: Defining dependency "gpudev" 00:02:25.798 Message: lib/gro: Defining dependency "gro" 00:02:25.798 Message: lib/gso: Defining dependency "gso" 00:02:25.798 Message: lib/ip_frag: Defining dependency "ip_frag" 00:02:25.798 Message: lib/jobstats: Defining dependency "jobstats" 00:02:25.798 Message: lib/latencystats: Defining dependency "latencystats" 00:02:25.798 Message: lib/lpm: Defining dependency "lpm" 00:02:25.798 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:25.798 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:02:25.798 Fetching value of define "__AVX512IFMA__" : (undefined) 00:02:25.798 Compiler for C supports arguments -mavx512f -mavx512dq -mavx512ifma: YES 00:02:25.798 Message: lib/member: Defining dependency "member" 00:02:25.798 Message: lib/pcapng: Defining dependency "pcapng" 00:02:25.798 Compiler for C supports arguments -Wno-cast-qual: YES 00:02:25.798 Message: lib/power: Defining dependency "power" 00:02:25.798 Message: lib/rawdev: Defining dependency "rawdev" 00:02:25.798 Message: lib/regexdev: Defining dependency "regexdev" 00:02:25.798 Message: lib/mldev: Defining dependency "mldev" 00:02:25.798 Message: lib/rib: Defining dependency "rib" 00:02:25.798 Message: lib/reorder: Defining dependency "reorder" 00:02:25.798 Message: lib/sched: Defining dependency "sched" 00:02:25.798 Message: lib/security: Defining dependency "security" 00:02:25.798 Message: lib/stack: Defining dependency "stack" 00:02:25.798 Has header "linux/userfaultfd.h" : YES 00:02:25.798 Has header "linux/vduse.h" : YES 00:02:25.798 Message: lib/vhost: Defining dependency "vhost" 00:02:25.798 Message: lib/ipsec: Defining dependency "ipsec" 00:02:25.798 Message: lib/pdcp: Defining dependency "pdcp" 00:02:25.798 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:25.798 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:02:25.798 Fetching value of define "__AVX512BW__" : 1 (cached) 00:02:25.798 Message: lib/fib: Defining dependency "fib" 00:02:25.798 Message: lib/port: Defining dependency "port" 00:02:25.798 Message: lib/pdump: Defining dependency "pdump" 00:02:25.798 Message: lib/table: Defining dependency "table" 00:02:25.798 Message: lib/pipeline: Defining dependency "pipeline" 00:02:25.798 Message: lib/graph: Defining dependency "graph" 00:02:25.798 Message: lib/node: Defining dependency "node" 00:02:25.798 Compiler for C supports arguments -Wno-format-truncation: YES (cached) 00:02:25.798 Message: drivers/bus/pci: Defining dependency "bus_pci" 00:02:25.798 Message: drivers/bus/vdev: Defining dependency "bus_vdev" 00:02:27.712 Message: drivers/mempool/ring: Defining dependency "mempool_ring" 00:02:27.712 Compiler for C supports arguments -Wno-sign-compare: YES 00:02:27.712 Compiler for C supports arguments -Wno-unused-value: YES 00:02:27.712 Compiler for C supports arguments -Wno-format: YES 00:02:27.712 Compiler for C supports arguments -Wno-format-security: YES 00:02:27.712 Compiler for C supports arguments -Wno-format-nonliteral: YES 00:02:27.712 Compiler for C supports arguments -Wno-strict-aliasing: YES 00:02:27.712 Compiler for C supports arguments -Wno-unused-but-set-variable: YES 00:02:27.712 Compiler for C supports arguments -Wno-unused-parameter: YES 00:02:27.712 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:27.712 Fetching value of define "__AVX512BW__" : 1 (cached) 00:02:27.712 Compiler for C supports arguments -mavx512f: YES (cached) 00:02:27.712 Compiler for C supports arguments -mavx512bw: YES (cached) 00:02:27.712 Compiler for C supports arguments -march=skylake-avx512: YES 00:02:27.712 Message: drivers/net/i40e: Defining dependency "net_i40e" 00:02:27.712 Has header "sys/epoll.h" : YES 00:02:27.712 Program doxygen found: YES (/usr/local/bin/doxygen) 00:02:27.712 Configuring doxy-api-html.conf using configuration 00:02:27.712 Configuring doxy-api-man.conf using configuration 00:02:27.712 Program mandb found: YES (/usr/bin/mandb) 00:02:27.712 Program sphinx-build found: NO 00:02:27.712 Configuring rte_build_config.h using configuration 00:02:27.712 Message: 00:02:27.712 ================= 00:02:27.712 Applications Enabled 00:02:27.712 ================= 00:02:27.712 00:02:27.712 apps: 00:02:27.712 dumpcap, graph, pdump, proc-info, test-acl, test-bbdev, test-cmdline, test-compress-perf, 00:02:27.712 test-crypto-perf, test-dma-perf, test-eventdev, test-fib, test-flow-perf, test-gpudev, test-mldev, test-pipeline, 00:02:27.712 test-pmd, test-regex, test-sad, test-security-perf, 00:02:27.712 00:02:27.712 Message: 00:02:27.712 ================= 00:02:27.712 Libraries Enabled 00:02:27.712 ================= 00:02:27.712 00:02:27.712 libs: 00:02:27.712 log, kvargs, telemetry, eal, ring, rcu, mempool, mbuf, 00:02:27.712 net, meter, ethdev, pci, cmdline, metrics, hash, timer, 00:02:27.712 acl, bbdev, bitratestats, bpf, cfgfile, compressdev, cryptodev, distributor, 00:02:27.712 dmadev, efd, eventdev, dispatcher, gpudev, gro, gso, ip_frag, 00:02:27.712 jobstats, latencystats, lpm, member, pcapng, power, rawdev, regexdev, 00:02:27.712 mldev, rib, reorder, sched, security, stack, vhost, ipsec, 00:02:27.712 pdcp, fib, port, pdump, table, pipeline, graph, node, 00:02:27.712 00:02:27.712 00:02:27.712 Message: 00:02:27.712 =============== 00:02:27.712 Drivers Enabled 00:02:27.712 =============== 00:02:27.712 00:02:27.712 common: 00:02:27.712 00:02:27.712 bus: 00:02:27.712 pci, vdev, 00:02:27.712 mempool: 00:02:27.712 ring, 00:02:27.712 dma: 00:02:27.712 00:02:27.712 net: 00:02:27.712 i40e, 00:02:27.712 raw: 00:02:27.712 00:02:27.712 crypto: 00:02:27.712 00:02:27.712 compress: 00:02:27.712 00:02:27.712 regex: 00:02:27.712 00:02:27.712 ml: 00:02:27.712 00:02:27.712 vdpa: 00:02:27.712 00:02:27.712 event: 00:02:27.712 00:02:27.712 baseband: 00:02:27.712 00:02:27.712 gpu: 00:02:27.712 00:02:27.712 00:02:27.712 Message: 00:02:27.712 ================= 00:02:27.712 Content Skipped 00:02:27.712 ================= 00:02:27.712 00:02:27.712 apps: 00:02:27.712 00:02:27.712 libs: 00:02:27.712 00:02:27.712 drivers: 00:02:27.712 common/cpt: not in enabled drivers build config 00:02:27.712 common/dpaax: not in enabled drivers build config 00:02:27.712 common/iavf: not in enabled drivers build config 00:02:27.712 common/idpf: not in enabled drivers build config 00:02:27.712 common/mvep: not in enabled drivers build config 00:02:27.712 common/octeontx: not in enabled drivers build config 00:02:27.712 bus/auxiliary: not in enabled drivers build config 00:02:27.712 bus/cdx: not in enabled drivers build config 00:02:27.712 bus/dpaa: not in enabled drivers build config 00:02:27.712 bus/fslmc: not in enabled drivers build config 00:02:27.712 bus/ifpga: not in enabled drivers build config 00:02:27.712 bus/platform: not in enabled drivers build config 00:02:27.712 bus/vmbus: not in enabled drivers build config 00:02:27.712 common/cnxk: not in enabled drivers build config 00:02:27.712 common/mlx5: not in enabled drivers build config 00:02:27.712 common/nfp: not in enabled drivers build config 00:02:27.712 common/qat: not in enabled drivers build config 00:02:27.712 common/sfc_efx: not in enabled drivers build config 00:02:27.712 mempool/bucket: not in enabled drivers build config 00:02:27.712 mempool/cnxk: not in enabled drivers build config 00:02:27.712 mempool/dpaa: not in enabled drivers build config 00:02:27.712 mempool/dpaa2: not in enabled drivers build config 00:02:27.712 mempool/octeontx: not in enabled drivers build config 00:02:27.712 mempool/stack: not in enabled drivers build config 00:02:27.712 dma/cnxk: not in enabled drivers build config 00:02:27.712 dma/dpaa: not in enabled drivers build config 00:02:27.712 dma/dpaa2: not in enabled drivers build config 00:02:27.712 dma/hisilicon: not in enabled drivers build config 00:02:27.712 dma/idxd: not in enabled drivers build config 00:02:27.712 dma/ioat: not in enabled drivers build config 00:02:27.712 dma/skeleton: not in enabled drivers build config 00:02:27.712 net/af_packet: not in enabled drivers build config 00:02:27.712 net/af_xdp: not in enabled drivers build config 00:02:27.713 net/ark: not in enabled drivers build config 00:02:27.713 net/atlantic: not in enabled drivers build config 00:02:27.713 net/avp: not in enabled drivers build config 00:02:27.713 net/axgbe: not in enabled drivers build config 00:02:27.713 net/bnx2x: not in enabled drivers build config 00:02:27.713 net/bnxt: not in enabled drivers build config 00:02:27.713 net/bonding: not in enabled drivers build config 00:02:27.713 net/cnxk: not in enabled drivers build config 00:02:27.713 net/cpfl: not in enabled drivers build config 00:02:27.713 net/cxgbe: not in enabled drivers build config 00:02:27.713 net/dpaa: not in enabled drivers build config 00:02:27.713 net/dpaa2: not in enabled drivers build config 00:02:27.713 net/e1000: not in enabled drivers build config 00:02:27.713 net/ena: not in enabled drivers build config 00:02:27.713 net/enetc: not in enabled drivers build config 00:02:27.713 net/enetfec: not in enabled drivers build config 00:02:27.713 net/enic: not in enabled drivers build config 00:02:27.713 net/failsafe: not in enabled drivers build config 00:02:27.713 net/fm10k: not in enabled drivers build config 00:02:27.713 net/gve: not in enabled drivers build config 00:02:27.713 net/hinic: not in enabled drivers build config 00:02:27.713 net/hns3: not in enabled drivers build config 00:02:27.713 net/iavf: not in enabled drivers build config 00:02:27.713 net/ice: not in enabled drivers build config 00:02:27.713 net/idpf: not in enabled drivers build config 00:02:27.713 net/igc: not in enabled drivers build config 00:02:27.713 net/ionic: not in enabled drivers build config 00:02:27.713 net/ipn3ke: not in enabled drivers build config 00:02:27.713 net/ixgbe: not in enabled drivers build config 00:02:27.713 net/mana: not in enabled drivers build config 00:02:27.713 net/memif: not in enabled drivers build config 00:02:27.713 net/mlx4: not in enabled drivers build config 00:02:27.713 net/mlx5: not in enabled drivers build config 00:02:27.713 net/mvneta: not in enabled drivers build config 00:02:27.713 net/mvpp2: not in enabled drivers build config 00:02:27.713 net/netvsc: not in enabled drivers build config 00:02:27.713 net/nfb: not in enabled drivers build config 00:02:27.713 net/nfp: not in enabled drivers build config 00:02:27.713 net/ngbe: not in enabled drivers build config 00:02:27.713 net/null: not in enabled drivers build config 00:02:27.713 net/octeontx: not in enabled drivers build config 00:02:27.713 net/octeon_ep: not in enabled drivers build config 00:02:27.713 net/pcap: not in enabled drivers build config 00:02:27.713 net/pfe: not in enabled drivers build config 00:02:27.713 net/qede: not in enabled drivers build config 00:02:27.713 net/ring: not in enabled drivers build config 00:02:27.713 net/sfc: not in enabled drivers build config 00:02:27.713 net/softnic: not in enabled drivers build config 00:02:27.713 net/tap: not in enabled drivers build config 00:02:27.713 net/thunderx: not in enabled drivers build config 00:02:27.713 net/txgbe: not in enabled drivers build config 00:02:27.713 net/vdev_netvsc: not in enabled drivers build config 00:02:27.713 net/vhost: not in enabled drivers build config 00:02:27.713 net/virtio: not in enabled drivers build config 00:02:27.713 net/vmxnet3: not in enabled drivers build config 00:02:27.713 raw/cnxk_bphy: not in enabled drivers build config 00:02:27.713 raw/cnxk_gpio: not in enabled drivers build config 00:02:27.713 raw/dpaa2_cmdif: not in enabled drivers build config 00:02:27.713 raw/ifpga: not in enabled drivers build config 00:02:27.713 raw/ntb: not in enabled drivers build config 00:02:27.713 raw/skeleton: not in enabled drivers build config 00:02:27.713 crypto/armv8: not in enabled drivers build config 00:02:27.713 crypto/bcmfs: not in enabled drivers build config 00:02:27.713 crypto/caam_jr: not in enabled drivers build config 00:02:27.713 crypto/ccp: not in enabled drivers build config 00:02:27.713 crypto/cnxk: not in enabled drivers build config 00:02:27.713 crypto/dpaa_sec: not in enabled drivers build config 00:02:27.713 crypto/dpaa2_sec: not in enabled drivers build config 00:02:27.713 crypto/ipsec_mb: not in enabled drivers build config 00:02:27.713 crypto/mlx5: not in enabled drivers build config 00:02:27.713 crypto/mvsam: not in enabled drivers build config 00:02:27.713 crypto/nitrox: not in enabled drivers build config 00:02:27.713 crypto/null: not in enabled drivers build config 00:02:27.713 crypto/octeontx: not in enabled drivers build config 00:02:27.713 crypto/openssl: not in enabled drivers build config 00:02:27.713 crypto/scheduler: not in enabled drivers build config 00:02:27.713 crypto/uadk: not in enabled drivers build config 00:02:27.713 crypto/virtio: not in enabled drivers build config 00:02:27.713 compress/isal: not in enabled drivers build config 00:02:27.713 compress/mlx5: not in enabled drivers build config 00:02:27.713 compress/octeontx: not in enabled drivers build config 00:02:27.713 compress/zlib: not in enabled drivers build config 00:02:27.713 regex/mlx5: not in enabled drivers build config 00:02:27.713 regex/cn9k: not in enabled drivers build config 00:02:27.713 ml/cnxk: not in enabled drivers build config 00:02:27.713 vdpa/ifc: not in enabled drivers build config 00:02:27.713 vdpa/mlx5: not in enabled drivers build config 00:02:27.713 vdpa/nfp: not in enabled drivers build config 00:02:27.713 vdpa/sfc: not in enabled drivers build config 00:02:27.713 event/cnxk: not in enabled drivers build config 00:02:27.713 event/dlb2: not in enabled drivers build config 00:02:27.713 event/dpaa: not in enabled drivers build config 00:02:27.713 event/dpaa2: not in enabled drivers build config 00:02:27.713 event/dsw: not in enabled drivers build config 00:02:27.713 event/opdl: not in enabled drivers build config 00:02:27.713 event/skeleton: not in enabled drivers build config 00:02:27.713 event/sw: not in enabled drivers build config 00:02:27.713 event/octeontx: not in enabled drivers build config 00:02:27.713 baseband/acc: not in enabled drivers build config 00:02:27.713 baseband/fpga_5gnr_fec: not in enabled drivers build config 00:02:27.713 baseband/fpga_lte_fec: not in enabled drivers build config 00:02:27.713 baseband/la12xx: not in enabled drivers build config 00:02:27.713 baseband/null: not in enabled drivers build config 00:02:27.713 baseband/turbo_sw: not in enabled drivers build config 00:02:27.713 gpu/cuda: not in enabled drivers build config 00:02:27.713 00:02:27.713 00:02:27.713 Build targets in project: 217 00:02:27.713 00:02:27.713 DPDK 23.11.0 00:02:27.713 00:02:27.713 User defined options 00:02:27.713 libdir : lib 00:02:27.713 prefix : /home/vagrant/spdk_repo/dpdk/build 00:02:27.713 c_args : -fPIC -g -fcommon -Werror -Wno-stringop-overflow 00:02:27.713 c_link_args : 00:02:27.713 enable_docs : false 00:02:27.713 enable_drivers: bus,bus/pci,bus/vdev,mempool/ring,net/i40e,net/i40e/base,power/acpi,power/amd_pstate,power/cppc,power/intel_pstate,power/intel_uncore,power/kvm_vm, 00:02:27.713 enable_kmods : false 00:02:27.713 machine : native 00:02:27.713 tests : false 00:02:27.713 00:02:27.713 Found ninja-1.11.1.git.kitware.jobserver-1 at /usr/local/bin/ninja 00:02:27.713 WARNING: Running the setup command as `meson [options]` instead of `meson setup [options]` is ambiguous and deprecated. 00:02:27.713 15:06:55 build_native_dpdk -- common/autobuild_common.sh@199 -- $ ninja -C /home/vagrant/spdk_repo/dpdk/build-tmp -j10 00:02:27.713 ninja: Entering directory `/home/vagrant/spdk_repo/dpdk/build-tmp' 00:02:27.713 [1/707] Compiling C object lib/librte_log.a.p/log_log_linux.c.o 00:02:27.972 [2/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_class.c.o 00:02:27.972 [3/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_debug.c.o 00:02:27.972 [4/707] Compiling C object lib/librte_kvargs.a.p/kvargs_rte_kvargs.c.o 00:02:27.972 [5/707] Linking static target lib/librte_kvargs.a 00:02:27.972 [6/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_errno.c.o 00:02:27.972 [7/707] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_data.c.o 00:02:27.972 [8/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hexdump.c.o 00:02:27.972 [9/707] Compiling C object lib/librte_log.a.p/log_log.c.o 00:02:27.972 [10/707] Linking static target lib/librte_log.a 00:02:27.972 [11/707] Generating lib/kvargs.sym_chk with a custom command (wrapped by meson to capture output) 00:02:27.972 [12/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_config.c.o 00:02:28.230 [13/707] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_legacy.c.o 00:02:28.230 [14/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_bus.c.o 00:02:28.230 [15/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dev.c.o 00:02:28.230 [16/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_devargs.c.o 00:02:28.230 [17/707] Generating lib/log.sym_chk with a custom command (wrapped by meson to capture output) 00:02:28.230 [18/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_interrupts.c.o 00:02:28.488 [19/707] Linking target lib/librte_log.so.24.0 00:02:28.488 [20/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_launch.c.o 00:02:28.488 [21/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_string_fns.c.o 00:02:28.488 [22/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_fbarray.c.o 00:02:28.488 [23/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_mcfg.c.o 00:02:28.488 [24/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_lcore.c.o 00:02:28.488 [25/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memalloc.c.o 00:02:28.748 [26/707] Generating symbol file lib/librte_log.so.24.0.p/librte_log.so.24.0.symbols 00:02:28.748 [27/707] Linking target lib/librte_kvargs.so.24.0 00:02:28.748 [28/707] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry.c.o 00:02:28.748 [29/707] Linking static target lib/librte_telemetry.a 00:02:28.748 [30/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_uuid.c.o 00:02:28.748 [31/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memzone.c.o 00:02:28.748 [32/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_tailqs.c.o 00:02:28.748 [33/707] Generating symbol file lib/librte_kvargs.so.24.0.p/librte_kvargs.so.24.0.symbols 00:02:28.748 [34/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_timer.c.o 00:02:28.748 [35/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_thread.c.o 00:02:29.018 [36/707] Compiling C object lib/librte_eal.a.p/eal_common_rte_reciprocal.c.o 00:02:29.018 [37/707] Compiling C object lib/librte_eal.a.p/eal_common_rte_version.c.o 00:02:29.018 [38/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memory.c.o 00:02:29.018 [39/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_cpuflags.c.o 00:02:29.018 [40/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hypervisor.c.o 00:02:29.018 [41/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_points.c.o 00:02:29.018 [42/707] Generating lib/telemetry.sym_chk with a custom command (wrapped by meson to capture output) 00:02:29.018 [43/707] Linking target lib/librte_telemetry.so.24.0 00:02:29.302 [44/707] Compiling C object lib/librte_eal.a.p/eal_common_malloc_elem.c.o 00:02:29.302 [45/707] Compiling C object lib/librte_eal.a.p/eal_common_rte_random.c.o 00:02:29.302 [46/707] Generating symbol file lib/librte_telemetry.so.24.0.p/librte_telemetry.so.24.0.symbols 00:02:29.302 [47/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_options.c.o 00:02:29.302 [48/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dynmem.c.o 00:02:29.302 [49/707] Compiling C object lib/librte_eal.a.p/eal_common_malloc_heap.c.o 00:02:29.302 [50/707] Compiling C object lib/librte_eal.a.p/eal_common_rte_malloc.c.o 00:02:29.302 [51/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace.c.o 00:02:29.561 [52/707] Compiling C object lib/librte_eal.a.p/eal_unix_eal_debug.c.o 00:02:29.561 [53/707] Compiling C object lib/librte_eal.a.p/eal_unix_eal_firmware.c.o 00:02:29.561 [54/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_ctf.c.o 00:02:29.561 [55/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_utils.c.o 00:02:29.561 [56/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_proc.c.o 00:02:29.561 [57/707] Compiling C object lib/librte_eal.a.p/eal_common_hotplug_mp.c.o 00:02:29.561 [58/707] Compiling C object lib/librte_eal.a.p/eal_common_rte_service.c.o 00:02:29.561 [59/707] Compiling C object lib/librte_eal.a.p/eal_common_rte_keepalive.c.o 00:02:29.819 [60/707] Compiling C object lib/librte_eal.a.p/eal_unix_eal_file.c.o 00:02:29.819 [61/707] Compiling C object lib/librte_eal.a.p/eal_unix_rte_thread.c.o 00:02:29.819 [62/707] Compiling C object lib/librte_eal.a.p/eal_linux_eal_cpuflags.c.o 00:02:29.819 [63/707] Compiling C object lib/librte_eal.a.p/eal_unix_eal_filesystem.c.o 00:02:29.819 [64/707] Compiling C object lib/librte_eal.a.p/eal_common_malloc_mp.c.o 00:02:29.819 [65/707] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_memory.c.o 00:02:29.819 [66/707] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_thread.c.o 00:02:29.819 [67/707] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_timer.c.o 00:02:29.819 [68/707] Compiling C object lib/librte_eal.a.p/eal_linux_eal_thread.c.o 00:02:30.077 [69/707] Compiling C object lib/librte_eal.a.p/eal_linux_eal_alarm.c.o 00:02:30.077 [70/707] Compiling C object lib/librte_eal.a.p/eal_linux_eal_lcore.c.o 00:02:30.077 [71/707] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio_mp_sync.c.o 00:02:30.077 [72/707] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cpuflags.c.o 00:02:30.077 [73/707] Compiling C object lib/librte_eal.a.p/eal_linux_eal_dev.c.o 00:02:30.077 [74/707] Compiling C object lib/librte_eal.a.p/eal_x86_rte_spinlock.c.o 00:02:30.077 [75/707] Compiling C object lib/librte_eal.a.p/eal_x86_rte_hypervisor.c.o 00:02:30.077 [76/707] Compiling C object lib/librte_eal.a.p/eal_linux_eal_hugepage_info.c.o 00:02:30.077 [77/707] Compiling C object lib/librte_eal.a.p/eal_linux_eal.c.o 00:02:30.077 [78/707] Compiling C object lib/librte_eal.a.p/eal_linux_eal_timer.c.o 00:02:30.335 [79/707] Compiling C object lib/librte_eal.a.p/eal_linux_eal_interrupts.c.o 00:02:30.335 [80/707] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memalloc.c.o 00:02:30.335 [81/707] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cycles.c.o 00:02:30.335 [82/707] Compiling C object lib/librte_eal.a.p/eal_x86_rte_power_intrinsics.c.o 00:02:30.594 [83/707] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memory.c.o 00:02:30.594 [84/707] Compiling C object lib/librte_ring.a.p/ring_rte_ring.c.o 00:02:30.594 [85/707] Linking static target lib/librte_ring.a 00:02:30.594 [86/707] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops.c.o 00:02:30.594 [87/707] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio.c.o 00:02:30.594 [88/707] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops_default.c.o 00:02:30.594 [89/707] Linking static target lib/librte_eal.a 00:02:30.853 [90/707] Generating lib/ring.sym_chk with a custom command (wrapped by meson to capture output) 00:02:30.853 [91/707] Compiling C object lib/librte_mempool.a.p/mempool_mempool_trace_points.c.o 00:02:30.853 [92/707] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_ptype.c.o 00:02:30.853 [93/707] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_pool_ops.c.o 00:02:30.853 [94/707] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool.c.o 00:02:30.853 [95/707] Linking static target lib/librte_mempool.a 00:02:31.112 [96/707] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_dyn.c.o 00:02:31.112 [97/707] Compiling C object lib/librte_rcu.a.p/rcu_rte_rcu_qsbr.c.o 00:02:31.112 [98/707] Linking static target lib/librte_rcu.a 00:02:31.112 [99/707] Compiling C object lib/net/libnet_crc_avx512_lib.a.p/net_crc_avx512.c.o 00:02:31.112 [100/707] Linking static target lib/net/libnet_crc_avx512_lib.a 00:02:31.112 [101/707] Compiling C object lib/librte_net.a.p/net_rte_ether.c.o 00:02:31.112 [102/707] Compiling C object lib/librte_net.a.p/net_rte_net_crc.c.o 00:02:31.112 [103/707] Compiling C object lib/librte_net.a.p/net_rte_arp.c.o 00:02:31.370 [104/707] Compiling C object lib/librte_net.a.p/net_rte_net.c.o 00:02:31.370 [105/707] Generating lib/rcu.sym_chk with a custom command (wrapped by meson to capture output) 00:02:31.370 [106/707] Compiling C object lib/librte_net.a.p/net_net_crc_sse.c.o 00:02:31.370 [107/707] Linking static target lib/librte_net.a 00:02:31.370 [108/707] Generating lib/mempool.sym_chk with a custom command (wrapped by meson to capture output) 00:02:31.627 [109/707] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf.c.o 00:02:31.628 [110/707] Linking static target lib/librte_mbuf.a 00:02:31.628 [111/707] Compiling C object lib/librte_meter.a.p/meter_rte_meter.c.o 00:02:31.628 [112/707] Linking static target lib/librte_meter.a 00:02:31.628 [113/707] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_profile.c.o 00:02:31.628 [114/707] Generating lib/net.sym_chk with a custom command (wrapped by meson to capture output) 00:02:31.628 [115/707] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_private.c.o 00:02:31.628 [116/707] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_driver.c.o 00:02:31.628 [117/707] Generating lib/meter.sym_chk with a custom command (wrapped by meson to capture output) 00:02:31.886 [118/707] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_class_eth.c.o 00:02:31.886 [119/707] Generating lib/mbuf.sym_chk with a custom command (wrapped by meson to capture output) 00:02:32.145 [120/707] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev_cman.c.o 00:02:32.145 [121/707] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_telemetry.c.o 00:02:32.403 [122/707] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev_telemetry.c.o 00:02:32.403 [123/707] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_common.c.o 00:02:32.403 [124/707] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_mtr.c.o 00:02:32.661 [125/707] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_trace_points.c.o 00:02:32.661 [126/707] Compiling C object lib/librte_pci.a.p/pci_rte_pci.c.o 00:02:32.661 [127/707] Linking static target lib/librte_pci.a 00:02:32.661 [128/707] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline.c.o 00:02:32.661 [129/707] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8079.c.o 00:02:32.661 [130/707] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8472.c.o 00:02:32.661 [131/707] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_tm.c.o 00:02:32.661 [132/707] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_cirbuf.c.o 00:02:32.661 [133/707] Generating lib/pci.sym_chk with a custom command (wrapped by meson to capture output) 00:02:32.661 [134/707] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse.c.o 00:02:32.920 [135/707] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_portlist.c.o 00:02:32.920 [136/707] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_num.c.o 00:02:32.920 [137/707] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_socket.c.o 00:02:32.920 [138/707] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_vt100.c.o 00:02:32.920 [139/707] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_string.c.o 00:02:32.920 [140/707] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_os_unix.c.o 00:02:32.920 [141/707] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_etheraddr.c.o 00:02:32.920 [142/707] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_ipaddr.c.o 00:02:32.920 [143/707] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_rdline.c.o 00:02:32.920 [144/707] Linking static target lib/librte_cmdline.a 00:02:32.920 [145/707] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8636.c.o 00:02:33.179 [146/707] Compiling C object lib/librte_metrics.a.p/metrics_rte_metrics_telemetry.c.o 00:02:33.179 [147/707] Compiling C object lib/librte_metrics.a.p/metrics_rte_metrics.c.o 00:02:33.179 [148/707] Linking static target lib/librte_metrics.a 00:02:33.179 [149/707] Compiling C object lib/librte_hash.a.p/hash_rte_fbk_hash.c.o 00:02:33.437 [150/707] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_flow.c.o 00:02:33.437 [151/707] Generating lib/metrics.sym_chk with a custom command (wrapped by meson to capture output) 00:02:33.695 [152/707] Compiling C object lib/librte_timer.a.p/timer_rte_timer.c.o 00:02:33.695 [153/707] Linking static target lib/librte_timer.a 00:02:33.695 [154/707] Compiling C object lib/librte_hash.a.p/hash_rte_thash.c.o 00:02:33.695 [155/707] Generating lib/cmdline.sym_chk with a custom command (wrapped by meson to capture output) 00:02:33.953 [156/707] Compiling C object lib/librte_acl.a.p/acl_acl_gen.c.o 00:02:33.954 [157/707] Generating lib/timer.sym_chk with a custom command (wrapped by meson to capture output) 00:02:33.954 [158/707] Compiling C object lib/librte_acl.a.p/acl_rte_acl.c.o 00:02:34.212 [159/707] Compiling C object lib/librte_acl.a.p/acl_tb_mem.c.o 00:02:34.212 [160/707] Compiling C object lib/librte_acl.a.p/acl_acl_run_scalar.c.o 00:02:34.470 [161/707] Compiling C object lib/librte_bitratestats.a.p/bitratestats_rte_bitrate.c.o 00:02:34.470 [162/707] Linking static target lib/librte_bitratestats.a 00:02:34.470 [163/707] Compiling C object lib/librte_bpf.a.p/bpf_bpf.c.o 00:02:34.728 [164/707] Generating lib/bitratestats.sym_chk with a custom command (wrapped by meson to capture output) 00:02:34.728 [165/707] Compiling C object lib/librte_acl.a.p/acl_acl_bld.c.o 00:02:34.728 [166/707] Compiling C object lib/librte_bbdev.a.p/bbdev_rte_bbdev.c.o 00:02:34.728 [167/707] Linking static target lib/librte_bbdev.a 00:02:34.985 [168/707] Compiling C object lib/librte_bpf.a.p/bpf_bpf_dump.c.o 00:02:34.985 [169/707] Compiling C object lib/librte_bpf.a.p/bpf_bpf_load.c.o 00:02:35.242 [170/707] Compiling C object lib/librte_hash.a.p/hash_rte_cuckoo_hash.c.o 00:02:35.242 [171/707] Compiling C object lib/librte_bpf.a.p/bpf_bpf_exec.c.o 00:02:35.242 [172/707] Linking static target lib/librte_hash.a 00:02:35.242 [173/707] Generating lib/bbdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:35.242 [174/707] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev.c.o 00:02:35.242 [175/707] Linking static target lib/librte_ethdev.a 00:02:35.242 [176/707] Compiling C object lib/acl/libavx2_tmp.a.p/acl_run_avx2.c.o 00:02:35.242 [177/707] Linking static target lib/acl/libavx2_tmp.a 00:02:35.242 [178/707] Compiling C object lib/librte_bpf.a.p/bpf_bpf_stub.c.o 00:02:35.500 [179/707] Compiling C object lib/librte_acl.a.p/acl_acl_run_sse.c.o 00:02:35.500 [180/707] Generating lib/eal.sym_chk with a custom command (wrapped by meson to capture output) 00:02:35.759 [181/707] Linking target lib/librte_eal.so.24.0 00:02:35.759 [182/707] Generating lib/hash.sym_chk with a custom command (wrapped by meson to capture output) 00:02:35.759 [183/707] Compiling C object lib/librte_cfgfile.a.p/cfgfile_rte_cfgfile.c.o 00:02:35.759 [184/707] Linking static target lib/librte_cfgfile.a 00:02:35.759 [185/707] Compiling C object lib/librte_bpf.a.p/bpf_bpf_pkt.c.o 00:02:35.759 [186/707] Compiling C object lib/librte_bpf.a.p/bpf_bpf_load_elf.c.o 00:02:35.759 [187/707] Generating symbol file lib/librte_eal.so.24.0.p/librte_eal.so.24.0.symbols 00:02:35.759 [188/707] Linking target lib/librte_ring.so.24.0 00:02:35.759 [189/707] Compiling C object lib/librte_bpf.a.p/bpf_bpf_convert.c.o 00:02:35.759 [190/707] Linking target lib/librte_meter.so.24.0 00:02:35.759 [191/707] Generating symbol file lib/librte_ring.so.24.0.p/librte_ring.so.24.0.symbols 00:02:36.018 [192/707] Linking target lib/librte_rcu.so.24.0 00:02:36.018 [193/707] Generating lib/cfgfile.sym_chk with a custom command (wrapped by meson to capture output) 00:02:36.018 [194/707] Generating symbol file lib/librte_meter.so.24.0.p/librte_meter.so.24.0.symbols 00:02:36.018 [195/707] Linking target lib/librte_mempool.so.24.0 00:02:36.018 [196/707] Generating symbol file lib/librte_rcu.so.24.0.p/librte_rcu.so.24.0.symbols 00:02:36.018 [197/707] Compiling C object lib/librte_bpf.a.p/bpf_bpf_validate.c.o 00:02:36.018 [198/707] Linking target lib/librte_pci.so.24.0 00:02:36.018 [199/707] Linking target lib/librte_cfgfile.so.24.0 00:02:36.018 [200/707] Linking target lib/librte_timer.so.24.0 00:02:36.018 [201/707] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev_pmd.c.o 00:02:36.018 [202/707] Generating symbol file lib/librte_mempool.so.24.0.p/librte_mempool.so.24.0.symbols 00:02:36.277 [203/707] Compiling C object lib/librte_bpf.a.p/bpf_bpf_jit_x86.c.o 00:02:36.277 [204/707] Generating symbol file lib/librte_pci.so.24.0.p/librte_pci.so.24.0.symbols 00:02:36.277 [205/707] Linking target lib/librte_mbuf.so.24.0 00:02:36.277 [206/707] Linking static target lib/librte_bpf.a 00:02:36.277 [207/707] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev.c.o 00:02:36.277 [208/707] Generating symbol file lib/librte_timer.so.24.0.p/librte_timer.so.24.0.symbols 00:02:36.277 [209/707] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_comp.c.o 00:02:36.277 [210/707] Linking static target lib/librte_compressdev.a 00:02:36.277 [211/707] Generating symbol file lib/librte_mbuf.so.24.0.p/librte_mbuf.so.24.0.symbols 00:02:36.277 [212/707] Linking target lib/librte_net.so.24.0 00:02:36.277 [213/707] Compiling C object lib/librte_acl.a.p/acl_acl_run_avx512.c.o 00:02:36.277 [214/707] Linking static target lib/librte_acl.a 00:02:36.535 [215/707] Generating symbol file lib/librte_net.so.24.0.p/librte_net.so.24.0.symbols 00:02:36.535 [216/707] Generating lib/bpf.sym_chk with a custom command (wrapped by meson to capture output) 00:02:36.535 [217/707] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_pmd.c.o 00:02:36.535 [218/707] Linking target lib/librte_hash.so.24.0 00:02:36.535 [219/707] Linking target lib/librte_cmdline.so.24.0 00:02:36.535 [220/707] Linking target lib/librte_bbdev.so.24.0 00:02:36.535 [221/707] Generating symbol file lib/librte_hash.so.24.0.p/librte_hash.so.24.0.symbols 00:02:36.535 [222/707] Compiling C object lib/librte_distributor.a.p/distributor_rte_distributor_single.c.o 00:02:36.535 [223/707] Generating lib/acl.sym_chk with a custom command (wrapped by meson to capture output) 00:02:36.793 [224/707] Generating lib/compressdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:36.793 [225/707] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_trace_points.c.o 00:02:36.793 [226/707] Linking target lib/librte_acl.so.24.0 00:02:36.793 [227/707] Linking target lib/librte_compressdev.so.24.0 00:02:36.793 [228/707] Compiling C object lib/librte_distributor.a.p/distributor_rte_distributor_match_sse.c.o 00:02:36.793 [229/707] Generating symbol file lib/librte_acl.so.24.0.p/librte_acl.so.24.0.symbols 00:02:36.793 [230/707] Compiling C object lib/librte_distributor.a.p/distributor_rte_distributor.c.o 00:02:36.793 [231/707] Linking static target lib/librte_distributor.a 00:02:37.052 [232/707] Compiling C object lib/librte_dmadev.a.p/dmadev_rte_dmadev_trace_points.c.o 00:02:37.052 [233/707] Compiling C object lib/librte_eventdev.a.p/eventdev_eventdev_private.c.o 00:02:37.052 [234/707] Generating lib/distributor.sym_chk with a custom command (wrapped by meson to capture output) 00:02:37.052 [235/707] Linking target lib/librte_distributor.so.24.0 00:02:37.052 [236/707] Compiling C object lib/librte_dmadev.a.p/dmadev_rte_dmadev.c.o 00:02:37.052 [237/707] Linking static target lib/librte_dmadev.a 00:02:37.311 [238/707] Compiling C object lib/librte_eventdev.a.p/eventdev_eventdev_trace_points.c.o 00:02:37.311 [239/707] Generating lib/dmadev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:37.571 [240/707] Linking target lib/librte_dmadev.so.24.0 00:02:37.571 [241/707] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_ring.c.o 00:02:37.571 [242/707] Generating symbol file lib/librte_dmadev.so.24.0.p/librte_dmadev.so.24.0.symbols 00:02:37.571 [243/707] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_dma_adapter.c.o 00:02:37.830 [244/707] Compiling C object lib/librte_efd.a.p/efd_rte_efd.c.o 00:02:37.830 [245/707] Linking static target lib/librte_efd.a 00:02:37.830 [246/707] Generating lib/efd.sym_chk with a custom command (wrapped by meson to capture output) 00:02:38.090 [247/707] Compiling C object lib/librte_cryptodev.a.p/cryptodev_rte_cryptodev.c.o 00:02:38.090 [248/707] Linking target lib/librte_efd.so.24.0 00:02:38.090 [249/707] Linking static target lib/librte_cryptodev.a 00:02:38.090 [250/707] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_crypto_adapter.c.o 00:02:38.090 [251/707] Compiling C object lib/librte_dispatcher.a.p/dispatcher_rte_dispatcher.c.o 00:02:38.090 [252/707] Linking static target lib/librte_dispatcher.a 00:02:38.090 [253/707] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_eth_tx_adapter.c.o 00:02:38.349 [254/707] Compiling C object lib/librte_gpudev.a.p/gpudev_gpudev.c.o 00:02:38.349 [255/707] Linking static target lib/librte_gpudev.a 00:02:38.349 [256/707] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_timer_adapter.c.o 00:02:38.349 [257/707] Generating lib/dispatcher.sym_chk with a custom command (wrapped by meson to capture output) 00:02:38.608 [258/707] Compiling C object lib/librte_gro.a.p/gro_gro_tcp4.c.o 00:02:38.608 [259/707] Compiling C object lib/librte_gro.a.p/gro_rte_gro.c.o 00:02:38.608 [260/707] Compiling C object lib/librte_gro.a.p/gro_gro_tcp6.c.o 00:02:38.868 [261/707] Compiling C object lib/librte_gro.a.p/gro_gro_udp4.c.o 00:02:38.868 [262/707] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_eventdev.c.o 00:02:38.868 [263/707] Compiling C object lib/librte_gro.a.p/gro_gro_vxlan_tcp4.c.o 00:02:39.127 [264/707] Compiling C object lib/librte_gso.a.p/gso_gso_tcp4.c.o 00:02:39.127 [265/707] Generating lib/gpudev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:39.127 [266/707] Generating lib/cryptodev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:39.127 [267/707] Compiling C object lib/librte_gro.a.p/gro_gro_vxlan_udp4.c.o 00:02:39.127 [268/707] Linking static target lib/librte_gro.a 00:02:39.127 [269/707] Linking target lib/librte_gpudev.so.24.0 00:02:39.127 [270/707] Linking target lib/librte_cryptodev.so.24.0 00:02:39.127 [271/707] Compiling C object lib/librte_gso.a.p/gso_gso_udp4.c.o 00:02:39.127 [272/707] Generating symbol file lib/librte_cryptodev.so.24.0.p/librte_cryptodev.so.24.0.symbols 00:02:39.127 [273/707] Compiling C object lib/librte_gso.a.p/gso_gso_tunnel_tcp4.c.o 00:02:39.127 [274/707] Generating lib/gro.sym_chk with a custom command (wrapped by meson to capture output) 00:02:39.387 [275/707] Compiling C object lib/librte_gso.a.p/gso_gso_common.c.o 00:02:39.387 [276/707] Compiling C object lib/librte_gso.a.p/gso_gso_tunnel_udp4.c.o 00:02:39.387 [277/707] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_eth_rx_adapter.c.o 00:02:39.387 [278/707] Linking static target lib/librte_eventdev.a 00:02:39.387 [279/707] Compiling C object lib/librte_gso.a.p/gso_rte_gso.c.o 00:02:39.387 [280/707] Linking static target lib/librte_gso.a 00:02:39.387 [281/707] Generating lib/ethdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:39.646 [282/707] Generating lib/gso.sym_chk with a custom command (wrapped by meson to capture output) 00:02:39.646 [283/707] Linking target lib/librte_ethdev.so.24.0 00:02:39.646 [284/707] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv6_reassembly.c.o 00:02:39.646 [285/707] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv4_reassembly.c.o 00:02:39.646 [286/707] Generating symbol file lib/librte_ethdev.so.24.0.p/librte_ethdev.so.24.0.symbols 00:02:39.646 [287/707] Compiling C object lib/librte_jobstats.a.p/jobstats_rte_jobstats.c.o 00:02:39.646 [288/707] Linking target lib/librte_metrics.so.24.0 00:02:39.646 [289/707] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv4_fragmentation.c.o 00:02:39.646 [290/707] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv6_fragmentation.c.o 00:02:39.646 [291/707] Linking target lib/librte_bpf.so.24.0 00:02:39.646 [292/707] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ip_frag_common.c.o 00:02:39.646 [293/707] Linking target lib/librte_gro.so.24.0 00:02:39.646 [294/707] Linking target lib/librte_gso.so.24.0 00:02:39.646 [295/707] Linking static target lib/librte_jobstats.a 00:02:39.905 [296/707] Generating symbol file lib/librte_metrics.so.24.0.p/librte_metrics.so.24.0.symbols 00:02:39.905 [297/707] Compiling C object lib/librte_ip_frag.a.p/ip_frag_ip_frag_internal.c.o 00:02:39.905 [298/707] Linking static target lib/librte_ip_frag.a 00:02:39.905 [299/707] Linking target lib/librte_bitratestats.so.24.0 00:02:39.905 [300/707] Generating symbol file lib/librte_bpf.so.24.0.p/librte_bpf.so.24.0.symbols 00:02:39.905 [301/707] Compiling C object lib/librte_latencystats.a.p/latencystats_rte_latencystats.c.o 00:02:39.905 [302/707] Generating lib/jobstats.sym_chk with a custom command (wrapped by meson to capture output) 00:02:39.905 [303/707] Linking static target lib/librte_latencystats.a 00:02:40.164 [304/707] Generating lib/ip_frag.sym_chk with a custom command (wrapped by meson to capture output) 00:02:40.164 [305/707] Linking target lib/librte_jobstats.so.24.0 00:02:40.164 [306/707] Linking target lib/librte_ip_frag.so.24.0 00:02:40.164 [307/707] Compiling C object lib/librte_member.a.p/member_rte_member.c.o 00:02:40.164 [308/707] Compiling C object lib/member/libsketch_avx512_tmp.a.p/rte_member_sketch_avx512.c.o 00:02:40.164 [309/707] Linking static target lib/member/libsketch_avx512_tmp.a 00:02:40.164 [310/707] Compiling C object lib/librte_lpm.a.p/lpm_rte_lpm.c.o 00:02:40.164 [311/707] Generating symbol file lib/librte_ip_frag.so.24.0.p/librte_ip_frag.so.24.0.symbols 00:02:40.164 [312/707] Generating lib/latencystats.sym_chk with a custom command (wrapped by meson to capture output) 00:02:40.164 [313/707] Linking target lib/librte_latencystats.so.24.0 00:02:40.164 [314/707] Compiling C object lib/librte_power.a.p/power_guest_channel.c.o 00:02:40.423 [315/707] Compiling C object lib/librte_power.a.p/power_power_common.c.o 00:02:40.423 [316/707] Compiling C object lib/librte_power.a.p/power_power_kvm_vm.c.o 00:02:40.423 [317/707] Compiling C object lib/librte_lpm.a.p/lpm_rte_lpm6.c.o 00:02:40.423 [318/707] Linking static target lib/librte_lpm.a 00:02:40.682 [319/707] Compiling C object lib/librte_power.a.p/power_power_acpi_cpufreq.c.o 00:02:40.682 [320/707] Compiling C object lib/librte_power.a.p/power_power_amd_pstate_cpufreq.c.o 00:02:40.682 [321/707] Compiling C object lib/librte_member.a.p/member_rte_member_ht.c.o 00:02:40.682 [322/707] Generating lib/lpm.sym_chk with a custom command (wrapped by meson to capture output) 00:02:40.682 [323/707] Linking target lib/librte_lpm.so.24.0 00:02:40.682 [324/707] Compiling C object lib/librte_pcapng.a.p/pcapng_rte_pcapng.c.o 00:02:40.682 [325/707] Linking static target lib/librte_pcapng.a 00:02:40.941 [326/707] Compiling C object lib/librte_power.a.p/power_power_cppc_cpufreq.c.o 00:02:40.941 [327/707] Compiling C object lib/librte_power.a.p/power_power_intel_uncore.c.o 00:02:40.941 [328/707] Generating symbol file lib/librte_lpm.so.24.0.p/librte_lpm.so.24.0.symbols 00:02:40.941 [329/707] Compiling C object lib/librte_member.a.p/member_rte_member_vbf.c.o 00:02:40.941 [330/707] Generating lib/pcapng.sym_chk with a custom command (wrapped by meson to capture output) 00:02:40.941 [331/707] Compiling C object lib/librte_power.a.p/power_rte_power.c.o 00:02:40.941 [332/707] Linking target lib/librte_pcapng.so.24.0 00:02:40.941 [333/707] Compiling C object lib/librte_power.a.p/power_rte_power_uncore.c.o 00:02:41.200 [334/707] Generating lib/eventdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:41.200 [335/707] Generating symbol file lib/librte_pcapng.so.24.0.p/librte_pcapng.so.24.0.symbols 00:02:41.200 [336/707] Linking target lib/librte_eventdev.so.24.0 00:02:41.200 [337/707] Compiling C object lib/librte_mldev.a.p/mldev_rte_mldev_pmd.c.o 00:02:41.200 [338/707] Generating symbol file lib/librte_eventdev.so.24.0.p/librte_eventdev.so.24.0.symbols 00:02:41.200 [339/707] Compiling C object lib/librte_power.a.p/power_power_pstate_cpufreq.c.o 00:02:41.200 [340/707] Linking target lib/librte_dispatcher.so.24.0 00:02:41.459 [341/707] Compiling C object lib/librte_power.a.p/power_rte_power_pmd_mgmt.c.o 00:02:41.459 [342/707] Compiling C object lib/librte_regexdev.a.p/regexdev_rte_regexdev.c.o 00:02:41.459 [343/707] Linking static target lib/librte_power.a 00:02:41.459 [344/707] Linking static target lib/librte_regexdev.a 00:02:41.459 [345/707] Compiling C object lib/librte_mldev.a.p/mldev_mldev_utils.c.o 00:02:41.459 [346/707] Compiling C object lib/librte_rawdev.a.p/rawdev_rte_rawdev.c.o 00:02:41.459 [347/707] Linking static target lib/librte_rawdev.a 00:02:41.459 [348/707] Compiling C object lib/librte_mldev.a.p/mldev_rte_mldev.c.o 00:02:41.459 [349/707] Compiling C object lib/librte_member.a.p/member_rte_member_sketch.c.o 00:02:41.717 [350/707] Compiling C object lib/librte_mldev.a.p/mldev_mldev_utils_scalar_bfloat16.c.o 00:02:41.717 [351/707] Linking static target lib/librte_member.a 00:02:41.717 [352/707] Compiling C object lib/librte_mldev.a.p/mldev_mldev_utils_scalar.c.o 00:02:41.717 [353/707] Linking static target lib/librte_mldev.a 00:02:41.717 [354/707] Generating lib/rawdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:41.717 [355/707] Linking target lib/librte_rawdev.so.24.0 00:02:41.717 [356/707] Generating lib/member.sym_chk with a custom command (wrapped by meson to capture output) 00:02:41.976 [357/707] Compiling C object lib/librte_rib.a.p/rib_rte_rib.c.o 00:02:41.977 [358/707] Linking target lib/librte_member.so.24.0 00:02:41.977 [359/707] Generating lib/power.sym_chk with a custom command (wrapped by meson to capture output) 00:02:41.977 [360/707] Compiling C object lib/librte_reorder.a.p/reorder_rte_reorder.c.o 00:02:41.977 [361/707] Linking static target lib/librte_reorder.a 00:02:41.977 [362/707] Linking target lib/librte_power.so.24.0 00:02:41.977 [363/707] Compiling C object lib/librte_sched.a.p/sched_rte_red.c.o 00:02:41.977 [364/707] Compiling C object lib/librte_sched.a.p/sched_rte_approx.c.o 00:02:41.977 [365/707] Generating lib/regexdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:41.977 [366/707] Linking target lib/librte_regexdev.so.24.0 00:02:42.235 [367/707] Compiling C object lib/librte_rib.a.p/rib_rte_rib6.c.o 00:02:42.235 [368/707] Linking static target lib/librte_rib.a 00:02:42.235 [369/707] Compiling C object lib/librte_vhost.a.p/vhost_fd_man.c.o 00:02:42.235 [370/707] Generating lib/reorder.sym_chk with a custom command (wrapped by meson to capture output) 00:02:42.235 [371/707] Compiling C object lib/librte_sched.a.p/sched_rte_pie.c.o 00:02:42.235 [372/707] Linking target lib/librte_reorder.so.24.0 00:02:42.235 [373/707] Compiling C object lib/librte_stack.a.p/stack_rte_stack_std.c.o 00:02:42.235 [374/707] Compiling C object lib/librte_stack.a.p/stack_rte_stack.c.o 00:02:42.235 [375/707] Compiling C object lib/librte_stack.a.p/stack_rte_stack_lf.c.o 00:02:42.235 [376/707] Linking static target lib/librte_stack.a 00:02:42.235 [377/707] Generating symbol file lib/librte_reorder.so.24.0.p/librte_reorder.so.24.0.symbols 00:02:42.494 [378/707] Compiling C object lib/librte_security.a.p/security_rte_security.c.o 00:02:42.495 [379/707] Linking static target lib/librte_security.a 00:02:42.495 [380/707] Generating lib/stack.sym_chk with a custom command (wrapped by meson to capture output) 00:02:42.495 [381/707] Linking target lib/librte_stack.so.24.0 00:02:42.495 [382/707] Generating lib/rib.sym_chk with a custom command (wrapped by meson to capture output) 00:02:42.495 [383/707] Linking target lib/librte_rib.so.24.0 00:02:42.753 [384/707] Generating symbol file lib/librte_rib.so.24.0.p/librte_rib.so.24.0.symbols 00:02:42.753 [385/707] Compiling C object lib/librte_vhost.a.p/vhost_vdpa.c.o 00:02:42.753 [386/707] Generating lib/mldev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:42.753 [387/707] Compiling C object lib/librte_vhost.a.p/vhost_iotlb.c.o 00:02:42.753 [388/707] Linking target lib/librte_mldev.so.24.0 00:02:42.753 [389/707] Generating lib/security.sym_chk with a custom command (wrapped by meson to capture output) 00:02:42.753 [390/707] Compiling C object lib/librte_vhost.a.p/vhost_socket.c.o 00:02:42.753 [391/707] Linking target lib/librte_security.so.24.0 00:02:43.012 [392/707] Compiling C object lib/librte_sched.a.p/sched_rte_sched.c.o 00:02:43.012 [393/707] Linking static target lib/librte_sched.a 00:02:43.012 [394/707] Generating symbol file lib/librte_security.so.24.0.p/librte_security.so.24.0.symbols 00:02:43.272 [395/707] Generating lib/sched.sym_chk with a custom command (wrapped by meson to capture output) 00:02:43.272 [396/707] Compiling C object lib/librte_vhost.a.p/vhost_virtio_net_ctrl.c.o 00:02:43.272 [397/707] Linking target lib/librte_sched.so.24.0 00:02:43.272 [398/707] Compiling C object lib/librte_vhost.a.p/vhost_vduse.c.o 00:02:43.272 [399/707] Generating symbol file lib/librte_sched.so.24.0.p/librte_sched.so.24.0.symbols 00:02:43.272 [400/707] Compiling C object lib/librte_vhost.a.p/vhost_vhost.c.o 00:02:43.531 [401/707] Compiling C object lib/librte_ipsec.a.p/ipsec_sa.c.o 00:02:43.531 [402/707] Compiling C object lib/librte_ipsec.a.p/ipsec_ses.c.o 00:02:43.531 [403/707] Compiling C object lib/librte_vhost.a.p/vhost_vhost_user.c.o 00:02:43.791 [404/707] Compiling C object lib/librte_ipsec.a.p/ipsec_ipsec_telemetry.c.o 00:02:43.791 [405/707] Compiling C object lib/librte_pdcp.a.p/pdcp_pdcp_cnt.c.o 00:02:43.791 [406/707] Compiling C object lib/librte_pdcp.a.p/pdcp_pdcp_crypto.c.o 00:02:44.058 [407/707] Compiling C object lib/librte_pdcp.a.p/pdcp_pdcp_ctrl_pdu.c.o 00:02:44.326 [408/707] Compiling C object lib/librte_pdcp.a.p/pdcp_pdcp_reorder.c.o 00:02:44.326 [409/707] Compiling C object lib/librte_fib.a.p/fib_rte_fib.c.o 00:02:44.326 [410/707] Compiling C object lib/librte_fib.a.p/fib_rte_fib6.c.o 00:02:44.326 [411/707] Compiling C object lib/librte_ipsec.a.p/ipsec_ipsec_sad.c.o 00:02:44.326 [412/707] Compiling C object lib/librte_ipsec.a.p/ipsec_esp_inb.c.o 00:02:44.326 [413/707] Compiling C object lib/librte_ipsec.a.p/ipsec_esp_outb.c.o 00:02:44.326 [414/707] Linking static target lib/librte_ipsec.a 00:02:44.326 [415/707] Compiling C object lib/librte_pdcp.a.p/pdcp_rte_pdcp.c.o 00:02:44.587 [416/707] Generating lib/ipsec.sym_chk with a custom command (wrapped by meson to capture output) 00:02:44.587 [417/707] Linking target lib/librte_ipsec.so.24.0 00:02:44.846 [418/707] Generating symbol file lib/librte_ipsec.so.24.0.p/librte_ipsec.so.24.0.symbols 00:02:44.846 [419/707] Compiling C object lib/librte_fib.a.p/fib_dir24_8_avx512.c.o 00:02:44.846 [420/707] Compiling C object lib/librte_fib.a.p/fib_trie_avx512.c.o 00:02:45.105 [421/707] Compiling C object lib/librte_fib.a.p/fib_trie.c.o 00:02:45.105 [422/707] Compiling C object lib/librte_port.a.p/port_rte_port_fd.c.o 00:02:45.106 [423/707] Compiling C object lib/librte_port.a.p/port_rte_port_ethdev.c.o 00:02:45.365 [424/707] Compiling C object lib/librte_fib.a.p/fib_dir24_8.c.o 00:02:45.365 [425/707] Linking static target lib/librte_fib.a 00:02:45.365 [426/707] Compiling C object lib/librte_pdcp.a.p/pdcp_pdcp_process.c.o 00:02:45.365 [427/707] Linking static target lib/librte_pdcp.a 00:02:45.365 [428/707] Compiling C object lib/librte_port.a.p/port_rte_port_frag.c.o 00:02:45.365 [429/707] Compiling C object lib/librte_port.a.p/port_rte_port_sched.c.o 00:02:45.365 [430/707] Compiling C object lib/librte_port.a.p/port_rte_port_ras.c.o 00:02:45.625 [431/707] Generating lib/fib.sym_chk with a custom command (wrapped by meson to capture output) 00:02:45.625 [432/707] Generating lib/pdcp.sym_chk with a custom command (wrapped by meson to capture output) 00:02:45.625 [433/707] Linking target lib/librte_fib.so.24.0 00:02:45.625 [434/707] Linking target lib/librte_pdcp.so.24.0 00:02:45.885 [435/707] Compiling C object lib/librte_port.a.p/port_rte_port_source_sink.c.o 00:02:45.885 [436/707] Compiling C object lib/librte_port.a.p/port_rte_port_sym_crypto.c.o 00:02:45.885 [437/707] Compiling C object lib/librte_table.a.p/table_rte_swx_keycmp.c.o 00:02:45.885 [438/707] Compiling C object lib/librte_port.a.p/port_rte_swx_port_fd.c.o 00:02:45.885 [439/707] Compiling C object lib/librte_port.a.p/port_rte_swx_port_ethdev.c.o 00:02:46.145 [440/707] Compiling C object lib/librte_port.a.p/port_rte_port_eventdev.c.o 00:02:46.145 [441/707] Compiling C object lib/librte_port.a.p/port_rte_swx_port_source_sink.c.o 00:02:46.405 [442/707] Compiling C object lib/librte_port.a.p/port_rte_swx_port_ring.c.o 00:02:46.405 [443/707] Compiling C object lib/librte_table.a.p/table_rte_swx_table_selector.c.o 00:02:46.405 [444/707] Compiling C object lib/librte_table.a.p/table_rte_swx_table_learner.c.o 00:02:46.405 [445/707] Compiling C object lib/librte_port.a.p/port_rte_port_ring.c.o 00:02:46.405 [446/707] Compiling C object lib/librte_table.a.p/table_rte_swx_table_wm.c.o 00:02:46.405 [447/707] Compiling C object lib/librte_table.a.p/table_rte_swx_table_em.c.o 00:02:46.405 [448/707] Linking static target lib/librte_port.a 00:02:46.664 [449/707] Compiling C object lib/librte_pdump.a.p/pdump_rte_pdump.c.o 00:02:46.664 [450/707] Linking static target lib/librte_pdump.a 00:02:46.664 [451/707] Compiling C object lib/librte_table.a.p/table_rte_table_array.c.o 00:02:46.664 [452/707] Compiling C object lib/librte_table.a.p/table_rte_table_acl.c.o 00:02:46.923 [453/707] Compiling C object lib/librte_table.a.p/table_rte_table_hash_cuckoo.c.o 00:02:46.923 [454/707] Generating lib/port.sym_chk with a custom command (wrapped by meson to capture output) 00:02:46.923 [455/707] Generating lib/pdump.sym_chk with a custom command (wrapped by meson to capture output) 00:02:46.923 [456/707] Linking target lib/librte_pdump.so.24.0 00:02:46.923 [457/707] Linking target lib/librte_port.so.24.0 00:02:47.182 [458/707] Generating symbol file lib/librte_port.so.24.0.p/librte_port.so.24.0.symbols 00:02:47.182 [459/707] Compiling C object lib/librte_table.a.p/table_rte_table_hash_key8.c.o 00:02:47.182 [460/707] Compiling C object lib/librte_table.a.p/table_rte_table_lpm.c.o 00:02:47.182 [461/707] Compiling C object lib/librte_table.a.p/table_rte_table_hash_ext.c.o 00:02:47.442 [462/707] Compiling C object lib/librte_table.a.p/table_rte_table_hash_key16.c.o 00:02:47.442 [463/707] Compiling C object lib/librte_table.a.p/table_rte_table_lpm_ipv6.c.o 00:02:47.442 [464/707] Compiling C object lib/librte_table.a.p/table_rte_table_stub.c.o 00:02:47.701 [465/707] Compiling C object lib/librte_vhost.a.p/vhost_vhost_crypto.c.o 00:02:47.701 [466/707] Compiling C object lib/librte_table.a.p/table_rte_table_hash_lru.c.o 00:02:47.701 [467/707] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_port_in_action.c.o 00:02:47.701 [468/707] Compiling C object lib/librte_table.a.p/table_rte_table_hash_key32.c.o 00:02:47.701 [469/707] Linking static target lib/librte_table.a 00:02:47.961 [470/707] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_pipeline.c.o 00:02:47.961 [471/707] Compiling C object lib/librte_graph.a.p/graph_node.c.o 00:02:48.221 [472/707] Generating lib/table.sym_chk with a custom command (wrapped by meson to capture output) 00:02:48.221 [473/707] Compiling C object lib/librte_graph.a.p/graph_graph.c.o 00:02:48.221 [474/707] Linking target lib/librte_table.so.24.0 00:02:48.221 [475/707] Generating symbol file lib/librte_table.so.24.0.p/librte_table.so.24.0.symbols 00:02:48.221 [476/707] Compiling C object lib/librte_graph.a.p/graph_graph_ops.c.o 00:02:48.480 [477/707] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_ipsec.c.o 00:02:48.480 [478/707] Compiling C object lib/librte_graph.a.p/graph_graph_debug.c.o 00:02:48.739 [479/707] Compiling C object lib/librte_graph.a.p/graph_graph_populate.c.o 00:02:48.739 [480/707] Compiling C object lib/librte_graph.a.p/graph_graph_stats.c.o 00:02:48.739 [481/707] Compiling C object lib/librte_graph.a.p/graph_rte_graph_worker.c.o 00:02:48.739 [482/707] Compiling C object lib/librte_graph.a.p/graph_graph_pcap.c.o 00:02:48.739 [483/707] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_ctl.c.o 00:02:48.998 [484/707] Compiling C object lib/librte_node.a.p/node_ethdev_ctrl.c.o 00:02:48.998 [485/707] Compiling C object lib/librte_node.a.p/node_ethdev_tx.c.o 00:02:49.257 [486/707] Compiling C object lib/librte_graph.a.p/graph_rte_graph_model_mcore_dispatch.c.o 00:02:49.257 [487/707] Linking static target lib/librte_graph.a 00:02:49.257 [488/707] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_pipeline_spec.c.o 00:02:49.257 [489/707] Compiling C object lib/librte_node.a.p/node_ethdev_rx.c.o 00:02:49.257 [490/707] Compiling C object lib/librte_node.a.p/node_ip4_local.c.o 00:02:49.516 [491/707] Generating lib/graph.sym_chk with a custom command (wrapped by meson to capture output) 00:02:49.516 [492/707] Compiling C object lib/librte_node.a.p/node_ip4_reassembly.c.o 00:02:49.516 [493/707] Linking target lib/librte_graph.so.24.0 00:02:49.775 [494/707] Compiling C object lib/librte_node.a.p/node_ip4_lookup.c.o 00:02:49.775 [495/707] Generating symbol file lib/librte_graph.so.24.0.p/librte_graph.so.24.0.symbols 00:02:49.775 [496/707] Compiling C object lib/librte_node.a.p/node_null.c.o 00:02:49.775 [497/707] Compiling C object lib/librte_node.a.p/node_ip6_lookup.c.o 00:02:49.775 [498/707] Compiling C object lib/librte_node.a.p/node_kernel_rx.c.o 00:02:50.034 [499/707] Compiling C object lib/librte_node.a.p/node_log.c.o 00:02:50.034 [500/707] Compiling C object lib/librte_node.a.p/node_ip4_rewrite.c.o 00:02:50.034 [501/707] Compiling C object lib/librte_node.a.p/node_kernel_tx.c.o 00:02:50.034 [502/707] Compiling C object lib/librte_node.a.p/node_ip6_rewrite.c.o 00:02:50.293 [503/707] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_params.c.o 00:02:50.293 [504/707] Compiling C object lib/librte_node.a.p/node_pkt_drop.c.o 00:02:50.552 [505/707] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common_uio.c.o 00:02:50.552 [506/707] Compiling C object lib/librte_node.a.p/node_udp4_input.c.o 00:02:50.552 [507/707] Compiling C object lib/librte_node.a.p/node_pkt_cls.c.o 00:02:50.552 [508/707] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev_params.c.o 00:02:50.552 [509/707] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common.c.o 00:02:50.552 [510/707] Linking static target lib/librte_node.a 00:02:50.552 [511/707] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci.c.o 00:02:50.812 [512/707] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_uio.c.o 00:02:50.812 [513/707] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_vfio.c.o 00:02:50.812 [514/707] Linking static target drivers/libtmp_rte_bus_pci.a 00:02:51.072 [515/707] Generating lib/node.sym_chk with a custom command (wrapped by meson to capture output) 00:02:51.072 [516/707] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev.c.o 00:02:51.072 [517/707] Linking static target drivers/libtmp_rte_bus_vdev.a 00:02:51.072 [518/707] Linking target lib/librte_node.so.24.0 00:02:51.072 [519/707] Generating drivers/rte_bus_pci.pmd.c with a custom command 00:02:51.072 [520/707] Compiling C object drivers/librte_bus_pci.a.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:02:51.072 [521/707] Linking static target drivers/librte_bus_pci.a 00:02:51.072 [522/707] Generating drivers/rte_bus_vdev.pmd.c with a custom command 00:02:51.072 [523/707] Compiling C object drivers/librte_bus_pci.so.24.0.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:02:51.072 [524/707] Compiling C object drivers/librte_bus_vdev.a.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:02:51.072 [525/707] Linking static target drivers/librte_bus_vdev.a 00:02:51.330 [526/707] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_dcb.c.o 00:02:51.330 [527/707] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_adminq.c.o 00:02:51.330 [528/707] Compiling C object drivers/librte_bus_vdev.so.24.0.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:02:51.330 [529/707] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_diag.c.o 00:02:51.330 [530/707] Generating drivers/rte_bus_vdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:51.330 [531/707] Linking target drivers/librte_bus_vdev.so.24.0 00:02:51.330 [532/707] Generating drivers/rte_bus_pci.sym_chk with a custom command (wrapped by meson to capture output) 00:02:51.589 [533/707] Linking target drivers/librte_bus_pci.so.24.0 00:02:51.589 [534/707] Compiling C object drivers/libtmp_rte_mempool_ring.a.p/mempool_ring_rte_mempool_ring.c.o 00:02:51.589 [535/707] Linking static target drivers/libtmp_rte_mempool_ring.a 00:02:51.589 [536/707] Generating symbol file drivers/librte_bus_vdev.so.24.0.p/librte_bus_vdev.so.24.0.symbols 00:02:51.589 [537/707] Generating symbol file drivers/librte_bus_pci.so.24.0.p/librte_bus_pci.so.24.0.symbols 00:02:51.589 [538/707] Generating drivers/rte_mempool_ring.pmd.c with a custom command 00:02:51.589 [539/707] Compiling C object drivers/librte_mempool_ring.a.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:02:51.589 [540/707] Linking static target drivers/librte_mempool_ring.a 00:02:51.589 [541/707] Compiling C object drivers/librte_mempool_ring.so.24.0.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:02:51.589 [542/707] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_hmc.c.o 00:02:51.589 [543/707] Linking target drivers/librte_mempool_ring.so.24.0 00:02:51.849 [544/707] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_lan_hmc.c.o 00:02:52.108 [545/707] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_nvm.c.o 00:02:52.371 [546/707] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_common.c.o 00:02:52.371 [547/707] Linking static target drivers/net/i40e/base/libi40e_base.a 00:02:52.938 [548/707] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_pf.c.o 00:02:53.197 [549/707] Compiling C object drivers/net/i40e/libi40e_avx512_lib.a.p/i40e_rxtx_vec_avx512.c.o 00:02:53.198 [550/707] Linking static target drivers/net/i40e/libi40e_avx512_lib.a 00:02:53.198 [551/707] Compiling C object drivers/net/i40e/libi40e_avx2_lib.a.p/i40e_rxtx_vec_avx2.c.o 00:02:53.198 [552/707] Linking static target drivers/net/i40e/libi40e_avx2_lib.a 00:02:53.458 [553/707] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_pipeline.c.o 00:02:53.458 [554/707] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_fdir.c.o 00:02:53.458 [555/707] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_tm.c.o 00:02:53.718 [556/707] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_flow.c.o 00:02:53.718 [557/707] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_vf_representor.c.o 00:02:53.977 [558/707] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_hash.c.o 00:02:53.977 [559/707] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_recycle_mbufs_vec_common.c.o 00:02:53.977 [560/707] Compiling C object app/dpdk-graph.p/graph_cli.c.o 00:02:54.236 [561/707] Compiling C object app/dpdk-dumpcap.p/dumpcap_main.c.o 00:02:54.236 [562/707] Compiling C object app/dpdk-graph.p/graph_conn.c.o 00:02:54.236 [563/707] Compiling C object app/dpdk-graph.p/graph_ethdev_rx.c.o 00:02:54.538 [564/707] Compiling C object app/dpdk-graph.p/graph_ethdev.c.o 00:02:54.806 [565/707] Compiling C object app/dpdk-graph.p/graph_ip4_route.c.o 00:02:54.806 [566/707] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_rxtx.c.o 00:02:54.806 [567/707] Compiling C object app/dpdk-graph.p/graph_ip6_route.c.o 00:02:54.806 [568/707] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_rxtx_vec_sse.c.o 00:02:54.806 [569/707] Compiling C object app/dpdk-graph.p/graph_graph.c.o 00:02:54.806 [570/707] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_rte_pmd_i40e.c.o 00:02:55.066 [571/707] Compiling C object app/dpdk-graph.p/graph_l3fwd.c.o 00:02:55.066 [572/707] Compiling C object app/dpdk-graph.p/graph_mempool.c.o 00:02:55.066 [573/707] Compiling C object app/dpdk-graph.p/graph_main.c.o 00:02:55.066 [574/707] Compiling C object app/dpdk-graph.p/graph_utils.c.o 00:02:55.326 [575/707] Compiling C object app/dpdk-graph.p/graph_neigh.c.o 00:02:55.585 [576/707] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_main.c.o 00:02:55.585 [577/707] Compiling C object app/dpdk-test-acl.p/test-acl_main.c.o 00:02:55.585 [578/707] Compiling C object app/dpdk-test-cmdline.p/test-cmdline_commands.c.o 00:02:55.585 [579/707] Compiling C object app/dpdk-test-cmdline.p/test-cmdline_cmdline_test.c.o 00:02:55.585 [580/707] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_ethdev.c.o 00:02:55.585 [581/707] Linking static target drivers/libtmp_rte_net_i40e.a 00:02:55.846 [582/707] Compiling C object app/dpdk-proc-info.p/proc-info_main.c.o 00:02:55.846 [583/707] Compiling C object app/dpdk-pdump.p/pdump_main.c.o 00:02:55.846 [584/707] Generating drivers/rte_net_i40e.pmd.c with a custom command 00:02:55.846 [585/707] Compiling C object drivers/librte_net_i40e.a.p/meson-generated_.._rte_net_i40e.pmd.c.o 00:02:55.846 [586/707] Linking static target drivers/librte_net_i40e.a 00:02:55.846 [587/707] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_test_bbdev.c.o 00:02:56.106 [588/707] Compiling C object drivers/librte_net_i40e.so.24.0.p/meson-generated_.._rte_net_i40e.pmd.c.o 00:02:56.106 [589/707] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_options_parse.c.o 00:02:56.106 [590/707] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_test_bbdev_vector.c.o 00:02:56.365 [591/707] Generating drivers/rte_net_i40e.sym_chk with a custom command (wrapped by meson to capture output) 00:02:56.365 [592/707] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_common.c.o 00:02:56.625 [593/707] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_main.c.o 00:02:56.625 [594/707] Linking target drivers/librte_net_i40e.so.24.0 00:02:56.625 [595/707] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_throughput.c.o 00:02:56.625 [596/707] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_verify.c.o 00:02:56.625 [597/707] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_cyclecount.c.o 00:02:56.885 [598/707] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_ops.c.o 00:02:56.885 [599/707] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_common.c.o 00:02:57.145 [600/707] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_options_parsing.c.o 00:02:57.145 [601/707] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_vector_parsing.c.o 00:02:57.145 [602/707] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_vectors.c.o 00:02:57.404 [603/707] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_latency.c.o 00:02:57.404 [604/707] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_pmd_cyclecount.c.o 00:02:57.404 [605/707] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_throughput.c.o 00:02:57.404 [606/707] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_evt_test.c.o 00:02:57.664 [607/707] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_parser.c.o 00:02:57.664 [608/707] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_verify.c.o 00:02:57.664 [609/707] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_evt_main.c.o 00:02:57.664 [610/707] Compiling C object app/dpdk-test-dma-perf.p/test-dma-perf_main.c.o 00:02:57.664 [611/707] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_main.c.o 00:02:57.924 [612/707] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_evt_options.c.o 00:02:57.924 [613/707] Compiling C object app/dpdk-test-dma-perf.p/test-dma-perf_benchmark.c.o 00:02:57.924 [614/707] Compiling C object lib/librte_vhost.a.p/vhost_virtio_net.c.o 00:02:57.924 [615/707] Linking static target lib/librte_vhost.a 00:02:58.183 [616/707] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_order_atq.c.o 00:02:58.183 [617/707] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_order_common.c.o 00:02:58.183 [618/707] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_order_queue.c.o 00:02:58.752 [619/707] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_pipeline_atq.c.o 00:02:58.752 [620/707] Generating lib/vhost.sym_chk with a custom command (wrapped by meson to capture output) 00:02:58.752 [621/707] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_perf_atq.c.o 00:02:59.012 [622/707] Linking target lib/librte_vhost.so.24.0 00:02:59.012 [623/707] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_pipeline_common.c.o 00:02:59.012 [624/707] Compiling C object app/dpdk-test-fib.p/test-fib_main.c.o 00:02:59.012 [625/707] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_perf_queue.c.o 00:02:59.012 [626/707] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_pipeline_queue.c.o 00:02:59.271 [627/707] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_flow_gen.c.o 00:02:59.271 [628/707] Compiling C object app/dpdk-test-mldev.p/test-mldev_ml_test.c.o 00:02:59.271 [629/707] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_items_gen.c.o 00:02:59.271 [630/707] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_actions_gen.c.o 00:02:59.530 [631/707] Compiling C object app/dpdk-test-mldev.p/test-mldev_ml_main.c.o 00:02:59.530 [632/707] Compiling C object app/dpdk-test-mldev.p/test-mldev_parser.c.o 00:02:59.530 [633/707] Compiling C object app/dpdk-test-gpudev.p/test-gpudev_main.c.o 00:02:59.530 [634/707] Compiling C object app/dpdk-test-mldev.p/test-mldev_ml_options.c.o 00:02:59.789 [635/707] Compiling C object app/dpdk-test-mldev.p/test-mldev_test_device_ops.c.o 00:02:59.789 [636/707] Compiling C object app/dpdk-test-mldev.p/test-mldev_test_common.c.o 00:02:59.789 [637/707] Compiling C object app/dpdk-test-mldev.p/test-mldev_test_model_common.c.o 00:02:59.789 [638/707] Compiling C object app/dpdk-test-mldev.p/test-mldev_test_model_ops.c.o 00:02:59.789 [639/707] Compiling C object app/dpdk-test-mldev.p/test-mldev_test_inference_ordered.c.o 00:03:00.048 [640/707] Compiling C object app/dpdk-test-mldev.p/test-mldev_test_inference_interleave.c.o 00:03:00.048 [641/707] Compiling C object app/dpdk-test-mldev.p/test-mldev_test_stats.c.o 00:03:00.048 [642/707] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_main.c.o 00:03:00.048 [643/707] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_config.c.o 00:03:00.308 [644/707] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_main.c.o 00:03:00.308 [645/707] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_init.c.o 00:03:00.308 [646/707] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_perf_common.c.o 00:03:00.568 [647/707] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_acl.c.o 00:03:00.568 [648/707] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_lpm.c.o 00:03:00.568 [649/707] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_lpm_ipv6.c.o 00:03:00.568 [650/707] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_hash.c.o 00:03:00.568 [651/707] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_stub.c.o 00:03:00.828 [652/707] Compiling C object app/dpdk-testpmd.p/test-pmd_5tswap.c.o 00:03:00.828 [653/707] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_test_bbdev_perf.c.o 00:03:00.828 [654/707] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_cman.c.o 00:03:01.086 [655/707] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_runtime.c.o 00:03:01.086 [656/707] Compiling C object app/dpdk-test-mldev.p/test-mldev_test_inference_common.c.o 00:03:01.345 [657/707] Compiling C object app/dpdk-testpmd.p/test-pmd_cmd_flex_item.c.o 00:03:01.345 [658/707] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_mtr.c.o 00:03:01.345 [659/707] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_tm.c.o 00:03:01.605 [660/707] Compiling C object app/dpdk-testpmd.p/test-pmd_flowgen.c.o 00:03:01.605 [661/707] Compiling C object app/dpdk-testpmd.p/test-pmd_iofwd.c.o 00:03:01.605 [662/707] Compiling C object app/dpdk-testpmd.p/test-pmd_icmpecho.c.o 00:03:01.865 [663/707] Compiling C object app/dpdk-testpmd.p/test-pmd_macfwd.c.o 00:03:01.865 [664/707] Compiling C object app/dpdk-testpmd.p/test-pmd_ieee1588fwd.c.o 00:03:01.865 [665/707] Compiling C object app/dpdk-testpmd.p/test-pmd_macswap.c.o 00:03:02.124 [666/707] Compiling C object app/dpdk-testpmd.p/test-pmd_recycle_mbufs.c.o 00:03:02.124 [667/707] Compiling C object app/dpdk-testpmd.p/test-pmd_rxonly.c.o 00:03:02.124 [668/707] Compiling C object app/dpdk-testpmd.p/test-pmd_csumonly.c.o 00:03:02.384 [669/707] Compiling C object app/dpdk-testpmd.p/test-pmd_shared_rxq_fwd.c.o 00:03:02.384 [670/707] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline.c.o 00:03:02.644 [671/707] Compiling C object app/dpdk-testpmd.p/test-pmd_parameters.c.o 00:03:02.905 [672/707] Compiling C object app/dpdk-testpmd.p/test-pmd_util.c.o 00:03:02.905 [673/707] Compiling C object app/dpdk-testpmd.p/test-pmd_bpf_cmd.c.o 00:03:03.165 [674/707] Compiling C object app/dpdk-testpmd.p/.._drivers_net_i40e_i40e_testpmd.c.o 00:03:03.426 [675/707] Compiling C object app/dpdk-testpmd.p/test-pmd_txonly.c.o 00:03:03.426 [676/707] Compiling C object app/dpdk-test-regex.p/test-regex_main.c.o 00:03:03.426 [677/707] Compiling C object app/dpdk-test-sad.p/test-sad_main.c.o 00:03:03.426 [678/707] Compiling C object app/dpdk-test-security-perf.p/test-security-perf_test_security_perf.c.o 00:03:03.426 [679/707] Compiling C object app/dpdk-testpmd.p/test-pmd_config.c.o 00:03:03.693 [680/707] Compiling C object app/dpdk-testpmd.p/test-pmd_noisy_vnf.c.o 00:03:03.693 [681/707] Compiling C object app/dpdk-testpmd.p/test-pmd_testpmd.c.o 00:03:03.693 [682/707] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_flow.c.o 00:03:03.962 [683/707] Compiling C object app/dpdk-test-security-perf.p/test_test_cryptodev_security_ipsec.c.o 00:03:04.222 [684/707] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_table_action.c.o 00:03:04.222 [685/707] Linking static target lib/librte_pipeline.a 00:03:04.481 [686/707] Linking target app/dpdk-test-cmdline 00:03:04.481 [687/707] Linking target app/dpdk-proc-info 00:03:04.481 [688/707] Linking target app/dpdk-test-acl 00:03:04.481 [689/707] Linking target app/dpdk-pdump 00:03:04.742 [690/707] Linking target app/dpdk-test-bbdev 00:03:04.742 [691/707] Linking target app/dpdk-dumpcap 00:03:04.742 [692/707] Linking target app/dpdk-test-compress-perf 00:03:04.742 [693/707] Linking target app/dpdk-graph 00:03:04.742 [694/707] Linking target app/dpdk-test-crypto-perf 00:03:05.002 [695/707] Linking target app/dpdk-test-dma-perf 00:03:05.002 [696/707] Linking target app/dpdk-test-fib 00:03:05.002 [697/707] Linking target app/dpdk-test-flow-perf 00:03:05.002 [698/707] Linking target app/dpdk-test-pipeline 00:03:05.002 [699/707] Linking target app/dpdk-test-gpudev 00:03:05.002 [700/707] Linking target app/dpdk-test-mldev 00:03:05.002 [701/707] Linking target app/dpdk-test-eventdev 00:03:05.002 [702/707] Linking target app/dpdk-testpmd 00:03:05.261 [703/707] Linking target app/dpdk-test-regex 00:03:05.261 [704/707] Linking target app/dpdk-test-security-perf 00:03:05.522 [705/707] Linking target app/dpdk-test-sad 00:03:10.816 [706/707] Generating lib/pipeline.sym_chk with a custom command (wrapped by meson to capture output) 00:03:10.816 [707/707] Linking target lib/librte_pipeline.so.24.0 00:03:10.816 15:07:38 build_native_dpdk -- common/autobuild_common.sh@201 -- $ uname -s 00:03:10.816 15:07:38 build_native_dpdk -- common/autobuild_common.sh@201 -- $ [[ Linux == \F\r\e\e\B\S\D ]] 00:03:10.816 15:07:38 build_native_dpdk -- common/autobuild_common.sh@214 -- $ ninja -C /home/vagrant/spdk_repo/dpdk/build-tmp -j10 install 00:03:10.816 ninja: Entering directory `/home/vagrant/spdk_repo/dpdk/build-tmp' 00:03:10.816 [0/1] Installing files. 00:03:10.816 Installing subdir /home/vagrant/spdk_repo/dpdk/examples to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples 00:03:10.816 Installing /home/vagrant/spdk_repo/dpdk/examples/bbdev_app/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bbdev_app 00:03:10.816 Installing /home/vagrant/spdk_repo/dpdk/examples/bbdev_app/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bbdev_app 00:03:10.816 Installing /home/vagrant/spdk_repo/dpdk/examples/bond/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bond 00:03:10.816 Installing /home/vagrant/spdk_repo/dpdk/examples/bond/commands.list to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bond 00:03:10.816 Installing /home/vagrant/spdk_repo/dpdk/examples/bond/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bond 00:03:10.816 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/README to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:03:10.816 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/dummy.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:03:10.816 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/t1.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:03:10.816 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/t2.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:03:10.816 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/t3.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:03:10.816 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:03:10.816 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/commands.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:03:10.816 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/commands.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/parse_obj_list.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/parse_obj_list.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/common/pkt_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/common/altivec/port_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common/altivec 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/common/neon/port_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common/neon 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/common/sse/port_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common/sse 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/distributor/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/distributor 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/distributor/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/distributor 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/dma/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/dma 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/dma/dmafwd.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/dma 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/ethapp.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/ethapp.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/lib/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/lib 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/lib/rte_ethtool.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/lib 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/lib/rte_ethtool.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/lib 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/pipeline_common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/pipeline_worker_generic.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/pipeline_worker_tx.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_dev_self_test.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_dev_self_test.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_aes.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_ccm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_cmac.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_ecdsa.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_gcm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_hmac.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_rsa.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_sha.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_tdes.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_xts.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_filtering/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_filtering 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_filtering/flow_blocks.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_filtering 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_filtering/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_filtering 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/helloworld/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/helloworld 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/helloworld/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/helloworld 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_fragmentation/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_fragmentation 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_fragmentation/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_fragmentation 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/action.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/action.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cli.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cli.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/conn.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/conn.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cryptodev.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cryptodev.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/link.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/link.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/mempool.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/mempool.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/parser.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/parser.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/pipeline.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/pipeline.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/swq.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/swq.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tap.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tap.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/thread.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/thread.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tmgr.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tmgr.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/firewall.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/flow.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/flow_crypto.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/l2fwd.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/route.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/route_ecmp.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/rss.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/tap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_reassembly/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_reassembly 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_reassembly/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_reassembly 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:10.817 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ep0.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ep1.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/esp.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/esp.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/event_helper.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/event_helper.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/flow.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/flow.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipip.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec-secgw.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec-secgw.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_lpm_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_process.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_worker.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_worker.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/parser.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/parser.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/rt.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sa.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sad.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sad.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sp4.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sp6.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/bypass_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/common_defs_secgw.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/data_rxtx.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/linux_test.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/load_env.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/pkttest.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/pkttest.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/run_test.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_3descbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_3descbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aescbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aescbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesctr_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesctr_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesgcm_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesgcm_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_ipv6opts.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_3descbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_3descbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aescbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aescbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesctr_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesctr_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesgcm_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesgcm_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_null_header_reconstruct.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipv4_multicast/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipv4_multicast 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/ipv4_multicast/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipv4_multicast 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/cat.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/cat.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/l2fwd-cat.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-crypto/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-crypto 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-crypto/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-crypto 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_common.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event_generic.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event_internal_port.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_poll.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_poll.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-jobstats/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-jobstats 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-jobstats/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-jobstats 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/shm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/shm.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/ka-agent/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive/ka-agent 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/ka-agent/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive/ka-agent 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-macsec/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-macsec 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-macsec/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-macsec 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd 00:03:10.818 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-graph/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-graph 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-graph/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-graph 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/perf_core.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/perf_core.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/em_default_v4.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/em_default_v6.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/em_route_parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_acl.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_acl.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_acl_scalar.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_altivec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_hlm.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_hlm_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_hlm_sse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_sequential.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event_generic.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event_internal_port.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_fib.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm_altivec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm_sse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_route.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_sse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/lpm_default_v4.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/lpm_default_v6.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/lpm_route_parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/link_status_interrupt/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/link_status_interrupt 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/link_status_interrupt/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/link_status_interrupt 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_client/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_client 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_client/client.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_client 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/args.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/args.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/init.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/init.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/shared/common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/shared 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/commands.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/commands.list to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/commands.list to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/mp_commands.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/mp_commands.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/symmetric_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/symmetric_mp 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/symmetric_mp/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/symmetric_mp 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/ntb/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ntb 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/ntb/commands.list to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ntb 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/ntb/ntb_fwd.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ntb 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/packet_ordering/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/packet_ordering 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/packet_ordering/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/packet_ordering 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/cli.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/cli.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/conn.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/conn.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/obj.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/obj.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/thread.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/thread.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/ethdev.io to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib_nexthop_group_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib_nexthop_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib_routing_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/hash_func.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.819 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/hash_func.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/ipsec.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/ipsec.io to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/ipsec.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/ipsec_sa.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_macswp.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_macswp.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_macswp_pcap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_pcap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/learner.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/learner.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/meter.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/meter.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/mirroring.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/mirroring.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/packet.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/pcap.io to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/recirculation.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/recirculation.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/registers.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/registers.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/rss.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/rss.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/selector.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/selector.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/selector.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/varbit.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/varbit.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan_pcap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan_table.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/ptpclient/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ptpclient 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/ptpclient/ptpclient.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ptpclient 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/rte_policer.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/rte_policer.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/app_thread.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/args.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/cfg_file.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/cfg_file.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/cmdline.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/init.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile_ov.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile_pie.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile_red.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/stats.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/rxtx_callbacks/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/rxtx_callbacks 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/rxtx_callbacks/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/rxtx_callbacks 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/efd_node/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/efd_node 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/efd_node/node.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/efd_node 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/efd_server/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/efd_server 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/efd_server/args.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/efd_server 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/efd_server/args.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/efd_server 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/efd_server/init.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/efd_server 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/efd_server/init.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/efd_server 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/efd_server/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/efd_server 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/shared/common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/shared 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/service_cores/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/service_cores 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/service_cores/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/service_cores 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/skeleton/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/skeleton 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/skeleton/basicfwd.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/skeleton 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/timer/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/timer 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/timer/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/timer 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/vdpa/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vdpa 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/vdpa/commands.list to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vdpa 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/vdpa/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vdpa 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/vdpa/vdpa_blk_compact.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vdpa 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/virtio_net.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/blk.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/blk_spec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:03:10.820 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/vhost_blk.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:03:10.821 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/vhost_blk.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:03:10.821 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/vhost_blk_compat.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:03:10.821 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_crypto/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_crypto 00:03:10.821 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_crypto/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_crypto 00:03:10.821 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:10.821 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_manager.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:10.821 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_manager.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:10.821 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_monitor.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:10.821 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_monitor.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:10.821 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:10.821 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/oob_monitor.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:10.821 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/oob_monitor_nop.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:10.821 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/oob_monitor_x86.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:10.821 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:10.821 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/parse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:10.821 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/power_manager.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:10.821 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/power_manager.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:10.821 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/vm_power_cli.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:10.821 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/vm_power_cli.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:10.821 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:03:10.821 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:03:10.821 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:03:10.821 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/parse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:03:10.821 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/vm_power_cli_guest.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:03:10.821 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/vm_power_cli_guest.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:03:10.821 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq 00:03:10.821 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq 00:03:10.821 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq_dcb/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq_dcb 00:03:10.821 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq_dcb/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq_dcb 00:03:10.821 Installing lib/librte_log.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_log.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_kvargs.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_kvargs.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_telemetry.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_telemetry.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_eal.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_eal.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_ring.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_ring.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_rcu.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_rcu.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_mempool.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_mempool.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_mbuf.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_mbuf.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_net.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_net.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_meter.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_meter.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_ethdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_ethdev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_pci.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_pci.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_cmdline.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_cmdline.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_metrics.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_metrics.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_hash.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_hash.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_timer.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_timer.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_acl.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_acl.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_bbdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_bbdev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_bitratestats.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_bitratestats.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_bpf.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_bpf.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_cfgfile.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_cfgfile.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_compressdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_compressdev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_cryptodev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_cryptodev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_distributor.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_distributor.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_dmadev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_dmadev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_efd.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_efd.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_eventdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_eventdev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_dispatcher.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_dispatcher.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_gpudev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_gpudev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_gro.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_gro.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_gso.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_gso.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_ip_frag.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_ip_frag.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_jobstats.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_jobstats.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_latencystats.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_latencystats.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_lpm.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_lpm.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_member.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_member.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.821 Installing lib/librte_pcapng.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_pcapng.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_power.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_power.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_rawdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_rawdev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_regexdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_regexdev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_mldev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_mldev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_rib.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_rib.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_reorder.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_reorder.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_sched.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_sched.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_security.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_security.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_stack.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_stack.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_vhost.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_vhost.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_ipsec.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_ipsec.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_pdcp.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_pdcp.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_fib.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_fib.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_port.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_port.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_pdump.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_pdump.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_table.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_table.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_pipeline.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_pipeline.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_graph.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:10.822 Installing lib/librte_graph.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:11.082 Installing lib/librte_node.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:11.082 Installing lib/librte_node.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:11.082 Installing drivers/librte_bus_pci.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:11.082 Installing drivers/librte_bus_pci.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0 00:03:11.082 Installing drivers/librte_bus_vdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:11.082 Installing drivers/librte_bus_vdev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0 00:03:11.082 Installing drivers/librte_mempool_ring.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:11.082 Installing drivers/librte_mempool_ring.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0 00:03:11.082 Installing drivers/librte_net_i40e.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:11.082 Installing drivers/librte_net_i40e.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0 00:03:11.082 Installing app/dpdk-dumpcap to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:11.082 Installing app/dpdk-graph to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:11.082 Installing app/dpdk-pdump to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:11.082 Installing app/dpdk-proc-info to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:11.082 Installing app/dpdk-test-acl to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:11.082 Installing app/dpdk-test-bbdev to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:11.082 Installing app/dpdk-test-cmdline to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:11.082 Installing app/dpdk-test-compress-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:11.082 Installing app/dpdk-test-crypto-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:11.082 Installing app/dpdk-test-dma-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:11.082 Installing app/dpdk-test-eventdev to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:11.082 Installing app/dpdk-test-fib to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:11.082 Installing app/dpdk-test-flow-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:11.082 Installing app/dpdk-test-gpudev to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:11.082 Installing app/dpdk-test-mldev to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:11.082 Installing app/dpdk-test-pipeline to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:11.082 Installing app/dpdk-testpmd to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:11.082 Installing app/dpdk-test-regex to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:11.082 Installing app/dpdk-test-sad to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:11.082 Installing app/dpdk-test-security-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/config/rte_config.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/log/rte_log.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/kvargs/rte_kvargs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/telemetry/rte_telemetry.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_atomic.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_byteorder.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_cpuflags.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_cycles.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_io.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_memcpy.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_pause.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_power_intrinsics.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_prefetch.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_rwlock.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_spinlock.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_vect.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_atomic.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_byteorder.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_cpuflags.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_cycles.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_io.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_memcpy.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_pause.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_power_intrinsics.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_prefetch.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_rtm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_rwlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_spinlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_vect.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_atomic_32.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_atomic_64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_byteorder_32.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_byteorder_64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_alarm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_bitmap.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_bitops.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_branch_prediction.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_bus.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_class.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_common.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_compat.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_debug.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_dev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_devargs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_eal.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_eal_memconfig.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_eal_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_errno.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_epoll.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_fbarray.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_hexdump.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_hypervisor.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_interrupts.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_keepalive.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_launch.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_lcore.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_lock_annotations.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_malloc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_mcslock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_memory.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_memzone.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_pci_dev_feature_defs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_pci_dev_features.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_per_lcore.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_pflock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_random.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_reciprocal.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_seqcount.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_seqlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_service.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_service_component.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.082 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_stdatomic.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_string_fns.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_tailq.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_thread.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_ticketlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_time.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_trace_point.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_trace_point_register.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_uuid.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_version.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_vfio.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/linux/include/rte_os.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_elem.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_c11_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_generic_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_hts.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_hts_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_peek.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_peek_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_peek_zc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_rts.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_rts_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/rcu/rte_rcu_qsbr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/mempool/rte_mempool.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/mempool/rte_mempool_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_ptype.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_pool_ops.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_dyn.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ip.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_tcp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_udp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_tls.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_dtls.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_esp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_sctp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_icmp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_arp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ether.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_macsec.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_vxlan.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_gre.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_gtp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_net.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_net_crc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.083 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_mpls.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_higig.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ecpri.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_pdcp_hdr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_geneve.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_l2tpv2.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ppp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ib.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/meter/rte_meter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_cman.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_dev_info.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_flow.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_flow_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_mtr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_mtr_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_tm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_tm_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_eth_ctrl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/pci/rte_pci.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_num.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_ipaddr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_etheraddr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_string.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_rdline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_vt100.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_socket.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_cirbuf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_portlist.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/metrics/rte_metrics.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/metrics/rte_metrics_telemetry.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_fbk_hash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_hash_crc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_hash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_jhash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_thash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_thash_gfni.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_arm64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_generic.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.343 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_sw.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_x86.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_thash_x86_gfni.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/timer/rte_timer.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/acl/rte_acl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/acl/rte_acl_osdep.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/bbdev/rte_bbdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/bbdev/rte_bbdev_pmd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/bbdev/rte_bbdev_op.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/bitratestats/rte_bitrate.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/bpf/bpf_def.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/bpf/rte_bpf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/bpf/rte_bpf_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/cfgfile/rte_cfgfile.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/compressdev/rte_compressdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/compressdev/rte_comp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_crypto.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_crypto_sym.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_crypto_asym.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/distributor/rte_distributor.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/dmadev/rte_dmadev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/dmadev/rte_dmadev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/efd/rte_efd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_crypto_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_dma_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_eth_rx_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_eth_tx_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_timer_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_eventdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_eventdev_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_eventdev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/dispatcher/rte_dispatcher.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/gpudev/rte_gpudev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/gro/rte_gro.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/gso/rte_gso.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/ip_frag/rte_ip_frag.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/jobstats/rte_jobstats.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/latencystats/rte_latencystats.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_altivec.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_neon.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_scalar.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_sse.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_sve.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/member/rte_member.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/pcapng/rte_pcapng.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_guest_channel.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_pmd_mgmt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_uncore.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/rawdev/rte_rawdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/rawdev/rte_rawdev_pmd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/regexdev/rte_regexdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/regexdev/rte_regexdev_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/regexdev/rte_regexdev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/mldev/rte_mldev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/mldev/rte_mldev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/rib/rte_rib.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/rib/rte_rib6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/reorder/rte_reorder.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_approx.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_red.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_sched.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_sched_common.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_pie.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/security/rte_security.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/security/rte_security_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_std.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf_generic.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf_c11.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf_stubs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vdpa.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vhost.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vhost_async.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vhost_crypto.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec_sa.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec_sad.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec_group.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/pdcp/rte_pdcp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/pdcp/rte_pdcp_group.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/fib/rte_fib.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/fib/rte_fib6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_fd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_frag.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_ras.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_sched.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_source_sink.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_sym_crypto.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_eventdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_fd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_source_sink.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.344 Installing /home/vagrant/spdk_repo/dpdk/lib/pdump/rte_pdump.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_lru.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_hash_func.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_em.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_learner.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_selector.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_wm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_acl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_array.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash_cuckoo.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash_func.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_lpm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_lpm_ipv6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_stub.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_lru_arm64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_lru_x86.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash_func_arm64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_pipeline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_port_in_action.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_table_action.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_swx_ipsec.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_swx_pipeline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_swx_extern.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_swx_ctl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/graph/rte_graph.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/graph/rte_graph_worker.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/graph/rte_graph_model_mcore_dispatch.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/graph/rte_graph_model_rtc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/graph/rte_graph_worker_common.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/node/rte_node_eth_api.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/node/rte_node_ip4_api.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/node/rte_node_ip6_api.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/lib/node/rte_node_udp4_input_api.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/drivers/bus/pci/rte_bus_pci.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/drivers/bus/vdev/rte_bus_vdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/drivers/net/i40e/rte_pmd_i40e.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/buildtools/dpdk-cmdline-gen.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-devbind.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-pmdinfo.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-telemetry.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-hugepages.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-rss-flows.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/build-tmp/rte_build_config.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/build-tmp/meson-private/libdpdk-libs.pc to /home/vagrant/spdk_repo/dpdk/build/lib/pkgconfig 00:03:11.345 Installing /home/vagrant/spdk_repo/dpdk/build-tmp/meson-private/libdpdk.pc to /home/vagrant/spdk_repo/dpdk/build/lib/pkgconfig 00:03:11.345 Installing symlink pointing to librte_log.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_log.so.24 00:03:11.345 Installing symlink pointing to librte_log.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_log.so 00:03:11.345 Installing symlink pointing to librte_kvargs.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_kvargs.so.24 00:03:11.345 Installing symlink pointing to librte_kvargs.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_kvargs.so 00:03:11.345 Installing symlink pointing to librte_telemetry.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_telemetry.so.24 00:03:11.345 Installing symlink pointing to librte_telemetry.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_telemetry.so 00:03:11.345 Installing symlink pointing to librte_eal.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eal.so.24 00:03:11.345 Installing symlink pointing to librte_eal.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eal.so 00:03:11.345 Installing symlink pointing to librte_ring.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ring.so.24 00:03:11.345 Installing symlink pointing to librte_ring.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ring.so 00:03:11.345 Installing symlink pointing to librte_rcu.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rcu.so.24 00:03:11.345 Installing symlink pointing to librte_rcu.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rcu.so 00:03:11.345 Installing symlink pointing to librte_mempool.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mempool.so.24 00:03:11.345 Installing symlink pointing to librte_mempool.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mempool.so 00:03:11.345 Installing symlink pointing to librte_mbuf.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mbuf.so.24 00:03:11.345 Installing symlink pointing to librte_mbuf.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mbuf.so 00:03:11.345 Installing symlink pointing to librte_net.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_net.so.24 00:03:11.345 Installing symlink pointing to librte_net.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_net.so 00:03:11.345 Installing symlink pointing to librte_meter.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_meter.so.24 00:03:11.345 Installing symlink pointing to librte_meter.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_meter.so 00:03:11.345 Installing symlink pointing to librte_ethdev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ethdev.so.24 00:03:11.345 Installing symlink pointing to librte_ethdev.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ethdev.so 00:03:11.345 Installing symlink pointing to librte_pci.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pci.so.24 00:03:11.345 Installing symlink pointing to librte_pci.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pci.so 00:03:11.345 Installing symlink pointing to librte_cmdline.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cmdline.so.24 00:03:11.345 Installing symlink pointing to librte_cmdline.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cmdline.so 00:03:11.345 Installing symlink pointing to librte_metrics.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_metrics.so.24 00:03:11.345 Installing symlink pointing to librte_metrics.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_metrics.so 00:03:11.345 Installing symlink pointing to librte_hash.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_hash.so.24 00:03:11.345 Installing symlink pointing to librte_hash.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_hash.so 00:03:11.345 Installing symlink pointing to librte_timer.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_timer.so.24 00:03:11.345 Installing symlink pointing to librte_timer.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_timer.so 00:03:11.345 Installing symlink pointing to librte_acl.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_acl.so.24 00:03:11.345 Installing symlink pointing to librte_acl.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_acl.so 00:03:11.345 Installing symlink pointing to librte_bbdev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bbdev.so.24 00:03:11.345 Installing symlink pointing to librte_bbdev.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bbdev.so 00:03:11.345 Installing symlink pointing to librte_bitratestats.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bitratestats.so.24 00:03:11.345 Installing symlink pointing to librte_bitratestats.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bitratestats.so 00:03:11.345 Installing symlink pointing to librte_bpf.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bpf.so.24 00:03:11.345 Installing symlink pointing to librte_bpf.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bpf.so 00:03:11.345 Installing symlink pointing to librte_cfgfile.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cfgfile.so.24 00:03:11.345 Installing symlink pointing to librte_cfgfile.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cfgfile.so 00:03:11.345 Installing symlink pointing to librte_compressdev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_compressdev.so.24 00:03:11.345 Installing symlink pointing to librte_compressdev.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_compressdev.so 00:03:11.345 Installing symlink pointing to librte_cryptodev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cryptodev.so.24 00:03:11.345 Installing symlink pointing to librte_cryptodev.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cryptodev.so 00:03:11.345 Installing symlink pointing to librte_distributor.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_distributor.so.24 00:03:11.345 Installing symlink pointing to librte_distributor.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_distributor.so 00:03:11.345 Installing symlink pointing to librte_dmadev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_dmadev.so.24 00:03:11.345 Installing symlink pointing to librte_dmadev.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_dmadev.so 00:03:11.345 Installing symlink pointing to librte_efd.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_efd.so.24 00:03:11.345 Installing symlink pointing to librte_efd.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_efd.so 00:03:11.345 Installing symlink pointing to librte_eventdev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eventdev.so.24 00:03:11.345 Installing symlink pointing to librte_eventdev.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eventdev.so 00:03:11.345 Installing symlink pointing to librte_dispatcher.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_dispatcher.so.24 00:03:11.345 Installing symlink pointing to librte_dispatcher.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_dispatcher.so 00:03:11.345 Installing symlink pointing to librte_gpudev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gpudev.so.24 00:03:11.345 Installing symlink pointing to librte_gpudev.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gpudev.so 00:03:11.345 Installing symlink pointing to librte_gro.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gro.so.24 00:03:11.346 Installing symlink pointing to librte_gro.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gro.so 00:03:11.346 Installing symlink pointing to librte_gso.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gso.so.24 00:03:11.346 Installing symlink pointing to librte_gso.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gso.so 00:03:11.346 Installing symlink pointing to librte_ip_frag.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ip_frag.so.24 00:03:11.346 Installing symlink pointing to librte_ip_frag.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ip_frag.so 00:03:11.346 Installing symlink pointing to librte_jobstats.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_jobstats.so.24 00:03:11.346 Installing symlink pointing to librte_jobstats.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_jobstats.so 00:03:11.346 Installing symlink pointing to librte_latencystats.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_latencystats.so.24 00:03:11.346 Installing symlink pointing to librte_latencystats.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_latencystats.so 00:03:11.346 Installing symlink pointing to librte_lpm.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_lpm.so.24 00:03:11.346 Installing symlink pointing to librte_lpm.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_lpm.so 00:03:11.346 Installing symlink pointing to librte_member.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_member.so.24 00:03:11.346 Installing symlink pointing to librte_member.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_member.so 00:03:11.346 Installing symlink pointing to librte_pcapng.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pcapng.so.24 00:03:11.346 Installing symlink pointing to librte_pcapng.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pcapng.so 00:03:11.346 Installing symlink pointing to librte_power.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_power.so.24 00:03:11.346 Installing symlink pointing to librte_power.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_power.so 00:03:11.346 Installing symlink pointing to librte_rawdev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rawdev.so.24 00:03:11.346 Installing symlink pointing to librte_rawdev.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rawdev.so 00:03:11.346 Installing symlink pointing to librte_regexdev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_regexdev.so.24 00:03:11.346 Installing symlink pointing to librte_regexdev.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_regexdev.so 00:03:11.346 Installing symlink pointing to librte_mldev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mldev.so.24 00:03:11.346 Installing symlink pointing to librte_mldev.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mldev.so 00:03:11.346 Installing symlink pointing to librte_rib.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rib.so.24 00:03:11.346 Installing symlink pointing to librte_rib.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rib.so 00:03:11.346 Installing symlink pointing to librte_reorder.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_reorder.so.24 00:03:11.346 Installing symlink pointing to librte_reorder.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_reorder.so 00:03:11.346 Installing symlink pointing to librte_sched.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_sched.so.24 00:03:11.346 Installing symlink pointing to librte_sched.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_sched.so 00:03:11.346 Installing symlink pointing to librte_security.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_security.so.24 00:03:11.346 Installing symlink pointing to librte_security.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_security.so 00:03:11.346 './librte_bus_pci.so' -> 'dpdk/pmds-24.0/librte_bus_pci.so' 00:03:11.346 './librte_bus_pci.so.24' -> 'dpdk/pmds-24.0/librte_bus_pci.so.24' 00:03:11.346 './librte_bus_pci.so.24.0' -> 'dpdk/pmds-24.0/librte_bus_pci.so.24.0' 00:03:11.346 './librte_bus_vdev.so' -> 'dpdk/pmds-24.0/librte_bus_vdev.so' 00:03:11.346 './librte_bus_vdev.so.24' -> 'dpdk/pmds-24.0/librte_bus_vdev.so.24' 00:03:11.346 './librte_bus_vdev.so.24.0' -> 'dpdk/pmds-24.0/librte_bus_vdev.so.24.0' 00:03:11.346 './librte_mempool_ring.so' -> 'dpdk/pmds-24.0/librte_mempool_ring.so' 00:03:11.346 './librte_mempool_ring.so.24' -> 'dpdk/pmds-24.0/librte_mempool_ring.so.24' 00:03:11.346 './librte_mempool_ring.so.24.0' -> 'dpdk/pmds-24.0/librte_mempool_ring.so.24.0' 00:03:11.346 './librte_net_i40e.so' -> 'dpdk/pmds-24.0/librte_net_i40e.so' 00:03:11.346 './librte_net_i40e.so.24' -> 'dpdk/pmds-24.0/librte_net_i40e.so.24' 00:03:11.346 './librte_net_i40e.so.24.0' -> 'dpdk/pmds-24.0/librte_net_i40e.so.24.0' 00:03:11.346 Installing symlink pointing to librte_stack.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_stack.so.24 00:03:11.346 Installing symlink pointing to librte_stack.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_stack.so 00:03:11.346 Installing symlink pointing to librte_vhost.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_vhost.so.24 00:03:11.346 Installing symlink pointing to librte_vhost.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_vhost.so 00:03:11.346 Installing symlink pointing to librte_ipsec.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ipsec.so.24 00:03:11.346 Installing symlink pointing to librte_ipsec.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ipsec.so 00:03:11.346 Installing symlink pointing to librte_pdcp.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pdcp.so.24 00:03:11.346 Installing symlink pointing to librte_pdcp.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pdcp.so 00:03:11.346 Installing symlink pointing to librte_fib.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_fib.so.24 00:03:11.346 Installing symlink pointing to librte_fib.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_fib.so 00:03:11.346 Installing symlink pointing to librte_port.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_port.so.24 00:03:11.346 Installing symlink pointing to librte_port.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_port.so 00:03:11.346 Installing symlink pointing to librte_pdump.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pdump.so.24 00:03:11.346 Installing symlink pointing to librte_pdump.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pdump.so 00:03:11.346 Installing symlink pointing to librte_table.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_table.so.24 00:03:11.346 Installing symlink pointing to librte_table.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_table.so 00:03:11.346 Installing symlink pointing to librte_pipeline.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pipeline.so.24 00:03:11.346 Installing symlink pointing to librte_pipeline.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pipeline.so 00:03:11.346 Installing symlink pointing to librte_graph.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_graph.so.24 00:03:11.346 Installing symlink pointing to librte_graph.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_graph.so 00:03:11.346 Installing symlink pointing to librte_node.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_node.so.24 00:03:11.346 Installing symlink pointing to librte_node.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_node.so 00:03:11.346 Installing symlink pointing to librte_bus_pci.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0/librte_bus_pci.so.24 00:03:11.346 Installing symlink pointing to librte_bus_pci.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0/librte_bus_pci.so 00:03:11.346 Installing symlink pointing to librte_bus_vdev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0/librte_bus_vdev.so.24 00:03:11.346 Installing symlink pointing to librte_bus_vdev.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0/librte_bus_vdev.so 00:03:11.346 Installing symlink pointing to librte_mempool_ring.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0/librte_mempool_ring.so.24 00:03:11.346 Installing symlink pointing to librte_mempool_ring.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0/librte_mempool_ring.so 00:03:11.346 Installing symlink pointing to librte_net_i40e.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0/librte_net_i40e.so.24 00:03:11.346 Installing symlink pointing to librte_net_i40e.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0/librte_net_i40e.so 00:03:11.346 Running custom install script '/bin/sh /home/vagrant/spdk_repo/dpdk/config/../buildtools/symlink-drivers-solibs.sh lib dpdk/pmds-24.0' 00:03:11.346 15:07:39 build_native_dpdk -- common/autobuild_common.sh@220 -- $ cat 00:03:11.346 15:07:39 build_native_dpdk -- common/autobuild_common.sh@225 -- $ cd /home/vagrant/spdk_repo/spdk 00:03:11.346 00:03:11.346 real 0m50.863s 00:03:11.346 user 5m12.915s 00:03:11.346 sys 0m59.080s 00:03:11.346 15:07:39 build_native_dpdk -- common/autotest_common.sh@1130 -- $ xtrace_disable 00:03:11.346 ************************************ 00:03:11.346 END TEST build_native_dpdk 00:03:11.346 ************************************ 00:03:11.346 15:07:39 build_native_dpdk -- common/autotest_common.sh@10 -- $ set +x 00:03:11.346 15:07:39 -- spdk/autobuild.sh@31 -- $ case "$SPDK_TEST_AUTOBUILD" in 00:03:11.346 15:07:39 -- spdk/autobuild.sh@47 -- $ [[ 0 -eq 1 ]] 00:03:11.346 15:07:39 -- spdk/autobuild.sh@51 -- $ [[ 0 -eq 1 ]] 00:03:11.346 15:07:39 -- spdk/autobuild.sh@55 -- $ [[ -n '' ]] 00:03:11.346 15:07:39 -- spdk/autobuild.sh@57 -- $ [[ 0 -eq 1 ]] 00:03:11.346 15:07:39 -- spdk/autobuild.sh@59 -- $ [[ 0 -eq 1 ]] 00:03:11.346 15:07:39 -- spdk/autobuild.sh@62 -- $ [[ 0 -eq 1 ]] 00:03:11.346 15:07:39 -- spdk/autobuild.sh@67 -- $ /home/vagrant/spdk_repo/spdk/configure --enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-asan --enable-coverage --with-ublk --with-raid5f --with-dpdk=/home/vagrant/spdk_repo/dpdk/build --with-shared 00:03:11.605 Using /home/vagrant/spdk_repo/dpdk/build/lib/pkgconfig for additional libs... 00:03:11.605 DPDK libraries: /home/vagrant/spdk_repo/dpdk/build/lib 00:03:11.605 DPDK includes: //home/vagrant/spdk_repo/dpdk/build/include 00:03:11.605 Using default SPDK env in /home/vagrant/spdk_repo/spdk/lib/env_dpdk 00:03:12.175 Using 'verbs' RDMA provider 00:03:28.000 Configuring ISA-L (logfile: /home/vagrant/spdk_repo/spdk/.spdk-isal.log)...done. 00:03:42.949 Configuring ISA-L-crypto (logfile: /home/vagrant/spdk_repo/spdk/.spdk-isal-crypto.log)...done. 00:03:43.208 Creating mk/config.mk...done. 00:03:43.208 Creating mk/cc.flags.mk...done. 00:03:43.208 Type 'make' to build. 00:03:43.208 15:08:11 -- spdk/autobuild.sh@70 -- $ run_test make make -j10 00:03:43.208 15:08:11 -- common/autotest_common.sh@1105 -- $ '[' 3 -le 1 ']' 00:03:43.208 15:08:11 -- common/autotest_common.sh@1111 -- $ xtrace_disable 00:03:43.208 15:08:11 -- common/autotest_common.sh@10 -- $ set +x 00:03:43.208 ************************************ 00:03:43.208 START TEST make 00:03:43.208 ************************************ 00:03:43.208 15:08:11 make -- common/autotest_common.sh@1129 -- $ make -j10 00:03:43.467 make[1]: Nothing to be done for 'all'. 00:04:30.164 CC lib/ut/ut.o 00:04:30.164 CC lib/log/log.o 00:04:30.164 CC lib/log/log_flags.o 00:04:30.164 CC lib/log/log_deprecated.o 00:04:30.164 CC lib/ut_mock/mock.o 00:04:30.164 LIB libspdk_ut.a 00:04:30.165 LIB libspdk_log.a 00:04:30.165 LIB libspdk_ut_mock.a 00:04:30.165 SO libspdk_ut.so.2.0 00:04:30.165 SO libspdk_ut_mock.so.6.0 00:04:30.165 SO libspdk_log.so.7.1 00:04:30.165 SYMLINK libspdk_ut.so 00:04:30.165 SYMLINK libspdk_ut_mock.so 00:04:30.165 SYMLINK libspdk_log.so 00:04:30.165 CC lib/dma/dma.o 00:04:30.165 CC lib/util/base64.o 00:04:30.165 CC lib/util/bit_array.o 00:04:30.165 CC lib/ioat/ioat.o 00:04:30.165 CC lib/util/cpuset.o 00:04:30.165 CC lib/util/crc32.o 00:04:30.165 CC lib/util/crc32c.o 00:04:30.165 CC lib/util/crc16.o 00:04:30.165 CXX lib/trace_parser/trace.o 00:04:30.165 CC lib/vfio_user/host/vfio_user_pci.o 00:04:30.165 CC lib/vfio_user/host/vfio_user.o 00:04:30.165 CC lib/util/crc32_ieee.o 00:04:30.165 CC lib/util/crc64.o 00:04:30.165 LIB libspdk_dma.a 00:04:30.165 CC lib/util/dif.o 00:04:30.165 SO libspdk_dma.so.5.0 00:04:30.165 CC lib/util/fd.o 00:04:30.165 CC lib/util/fd_group.o 00:04:30.165 CC lib/util/file.o 00:04:30.165 CC lib/util/hexlify.o 00:04:30.165 SYMLINK libspdk_dma.so 00:04:30.165 CC lib/util/iov.o 00:04:30.165 LIB libspdk_ioat.a 00:04:30.165 SO libspdk_ioat.so.7.0 00:04:30.165 CC lib/util/math.o 00:04:30.165 CC lib/util/net.o 00:04:30.165 LIB libspdk_vfio_user.a 00:04:30.165 SYMLINK libspdk_ioat.so 00:04:30.165 CC lib/util/pipe.o 00:04:30.165 SO libspdk_vfio_user.so.5.0 00:04:30.165 CC lib/util/strerror_tls.o 00:04:30.165 CC lib/util/string.o 00:04:30.165 SYMLINK libspdk_vfio_user.so 00:04:30.165 CC lib/util/uuid.o 00:04:30.165 CC lib/util/xor.o 00:04:30.165 CC lib/util/zipf.o 00:04:30.165 CC lib/util/md5.o 00:04:30.165 LIB libspdk_util.a 00:04:30.165 LIB libspdk_trace_parser.a 00:04:30.165 SO libspdk_util.so.10.1 00:04:30.165 SO libspdk_trace_parser.so.6.0 00:04:30.165 SYMLINK libspdk_util.so 00:04:30.165 SYMLINK libspdk_trace_parser.so 00:04:30.165 CC lib/json/json_util.o 00:04:30.165 CC lib/json/json_write.o 00:04:30.165 CC lib/rdma_utils/rdma_utils.o 00:04:30.165 CC lib/idxd/idxd.o 00:04:30.165 CC lib/idxd/idxd_kernel.o 00:04:30.165 CC lib/json/json_parse.o 00:04:30.165 CC lib/idxd/idxd_user.o 00:04:30.165 CC lib/conf/conf.o 00:04:30.165 CC lib/vmd/vmd.o 00:04:30.165 CC lib/env_dpdk/env.o 00:04:30.165 CC lib/vmd/led.o 00:04:30.165 LIB libspdk_conf.a 00:04:30.165 CC lib/env_dpdk/memory.o 00:04:30.165 CC lib/env_dpdk/pci.o 00:04:30.165 SO libspdk_conf.so.6.0 00:04:30.165 LIB libspdk_json.a 00:04:30.165 CC lib/env_dpdk/init.o 00:04:30.165 LIB libspdk_rdma_utils.a 00:04:30.165 SYMLINK libspdk_conf.so 00:04:30.165 CC lib/env_dpdk/threads.o 00:04:30.165 CC lib/env_dpdk/pci_ioat.o 00:04:30.165 SO libspdk_json.so.6.0 00:04:30.165 SO libspdk_rdma_utils.so.1.0 00:04:30.165 SYMLINK libspdk_json.so 00:04:30.165 SYMLINK libspdk_rdma_utils.so 00:04:30.165 CC lib/env_dpdk/pci_virtio.o 00:04:30.165 CC lib/env_dpdk/pci_vmd.o 00:04:30.165 CC lib/env_dpdk/pci_idxd.o 00:04:30.165 CC lib/jsonrpc/jsonrpc_server.o 00:04:30.165 CC lib/env_dpdk/pci_event.o 00:04:30.165 CC lib/rdma_provider/common.o 00:04:30.165 CC lib/rdma_provider/rdma_provider_verbs.o 00:04:30.165 CC lib/env_dpdk/sigbus_handler.o 00:04:30.165 CC lib/env_dpdk/pci_dpdk.o 00:04:30.165 LIB libspdk_idxd.a 00:04:30.165 SO libspdk_idxd.so.12.1 00:04:30.165 CC lib/jsonrpc/jsonrpc_server_tcp.o 00:04:30.165 LIB libspdk_vmd.a 00:04:30.165 CC lib/jsonrpc/jsonrpc_client.o 00:04:30.165 CC lib/jsonrpc/jsonrpc_client_tcp.o 00:04:30.165 SO libspdk_vmd.so.6.0 00:04:30.165 CC lib/env_dpdk/pci_dpdk_2207.o 00:04:30.165 SYMLINK libspdk_idxd.so 00:04:30.165 CC lib/env_dpdk/pci_dpdk_2211.o 00:04:30.165 LIB libspdk_rdma_provider.a 00:04:30.165 SYMLINK libspdk_vmd.so 00:04:30.165 SO libspdk_rdma_provider.so.7.0 00:04:30.165 SYMLINK libspdk_rdma_provider.so 00:04:30.165 LIB libspdk_jsonrpc.a 00:04:30.165 SO libspdk_jsonrpc.so.6.0 00:04:30.165 SYMLINK libspdk_jsonrpc.so 00:04:30.165 CC lib/rpc/rpc.o 00:04:30.165 LIB libspdk_env_dpdk.a 00:04:30.165 SO libspdk_env_dpdk.so.15.1 00:04:30.165 LIB libspdk_rpc.a 00:04:30.165 SO libspdk_rpc.so.6.0 00:04:30.165 SYMLINK libspdk_env_dpdk.so 00:04:30.165 SYMLINK libspdk_rpc.so 00:04:30.165 CC lib/notify/notify.o 00:04:30.165 CC lib/notify/notify_rpc.o 00:04:30.165 CC lib/trace/trace.o 00:04:30.165 CC lib/trace/trace_rpc.o 00:04:30.165 CC lib/trace/trace_flags.o 00:04:30.165 CC lib/keyring/keyring.o 00:04:30.165 CC lib/keyring/keyring_rpc.o 00:04:30.165 LIB libspdk_notify.a 00:04:30.165 SO libspdk_notify.so.6.0 00:04:30.165 SYMLINK libspdk_notify.so 00:04:30.165 LIB libspdk_keyring.a 00:04:30.165 SO libspdk_keyring.so.2.0 00:04:30.165 LIB libspdk_trace.a 00:04:30.165 SYMLINK libspdk_keyring.so 00:04:30.165 SO libspdk_trace.so.11.0 00:04:30.165 SYMLINK libspdk_trace.so 00:04:30.425 CC lib/thread/thread.o 00:04:30.425 CC lib/thread/iobuf.o 00:04:30.425 CC lib/sock/sock.o 00:04:30.425 CC lib/sock/sock_rpc.o 00:04:30.684 LIB libspdk_sock.a 00:04:30.943 SO libspdk_sock.so.10.0 00:04:30.943 SYMLINK libspdk_sock.so 00:04:31.511 CC lib/nvme/nvme_ctrlr_cmd.o 00:04:31.511 CC lib/nvme/nvme_ctrlr.o 00:04:31.511 CC lib/nvme/nvme_fabric.o 00:04:31.511 CC lib/nvme/nvme_ns_cmd.o 00:04:31.511 CC lib/nvme/nvme_ns.o 00:04:31.511 CC lib/nvme/nvme_pcie_common.o 00:04:31.511 CC lib/nvme/nvme_pcie.o 00:04:31.511 CC lib/nvme/nvme_qpair.o 00:04:31.511 CC lib/nvme/nvme.o 00:04:32.078 LIB libspdk_thread.a 00:04:32.078 CC lib/nvme/nvme_quirks.o 00:04:32.078 CC lib/nvme/nvme_transport.o 00:04:32.078 SO libspdk_thread.so.11.0 00:04:32.078 CC lib/nvme/nvme_discovery.o 00:04:32.078 SYMLINK libspdk_thread.so 00:04:32.078 CC lib/nvme/nvme_ctrlr_ocssd_cmd.o 00:04:32.078 CC lib/nvme/nvme_ns_ocssd_cmd.o 00:04:32.337 CC lib/nvme/nvme_tcp.o 00:04:32.337 CC lib/nvme/nvme_opal.o 00:04:32.337 CC lib/nvme/nvme_io_msg.o 00:04:32.337 CC lib/nvme/nvme_poll_group.o 00:04:32.596 CC lib/nvme/nvme_zns.o 00:04:32.596 CC lib/nvme/nvme_stubs.o 00:04:32.596 CC lib/nvme/nvme_auth.o 00:04:32.854 CC lib/nvme/nvme_cuse.o 00:04:32.854 CC lib/nvme/nvme_rdma.o 00:04:33.113 CC lib/accel/accel.o 00:04:33.113 CC lib/blob/blobstore.o 00:04:33.113 CC lib/blob/request.o 00:04:33.113 CC lib/blob/zeroes.o 00:04:33.113 CC lib/accel/accel_rpc.o 00:04:33.371 CC lib/accel/accel_sw.o 00:04:33.371 CC lib/blob/blob_bs_dev.o 00:04:33.630 CC lib/init/json_config.o 00:04:33.630 CC lib/init/subsystem.o 00:04:33.630 CC lib/init/subsystem_rpc.o 00:04:33.630 CC lib/virtio/virtio.o 00:04:33.889 CC lib/fsdev/fsdev.o 00:04:33.889 CC lib/init/rpc.o 00:04:33.889 CC lib/fsdev/fsdev_io.o 00:04:33.889 CC lib/fsdev/fsdev_rpc.o 00:04:33.889 CC lib/virtio/virtio_vhost_user.o 00:04:33.889 CC lib/virtio/virtio_vfio_user.o 00:04:33.889 LIB libspdk_init.a 00:04:33.889 CC lib/virtio/virtio_pci.o 00:04:33.889 SO libspdk_init.so.6.0 00:04:34.148 SYMLINK libspdk_init.so 00:04:34.148 LIB libspdk_accel.a 00:04:34.148 SO libspdk_accel.so.16.0 00:04:34.148 CC lib/event/app.o 00:04:34.148 CC lib/event/reactor.o 00:04:34.148 CC lib/event/log_rpc.o 00:04:34.148 LIB libspdk_nvme.a 00:04:34.148 LIB libspdk_virtio.a 00:04:34.148 CC lib/event/app_rpc.o 00:04:34.148 CC lib/event/scheduler_static.o 00:04:34.406 SO libspdk_virtio.so.7.0 00:04:34.406 SYMLINK libspdk_accel.so 00:04:34.406 SYMLINK libspdk_virtio.so 00:04:34.406 LIB libspdk_fsdev.a 00:04:34.406 SO libspdk_nvme.so.15.0 00:04:34.406 SO libspdk_fsdev.so.2.0 00:04:34.664 SYMLINK libspdk_fsdev.so 00:04:34.664 CC lib/bdev/bdev.o 00:04:34.664 CC lib/bdev/part.o 00:04:34.664 CC lib/bdev/bdev_rpc.o 00:04:34.664 CC lib/bdev/bdev_zone.o 00:04:34.664 CC lib/bdev/scsi_nvme.o 00:04:34.664 SYMLINK libspdk_nvme.so 00:04:34.664 CC lib/fuse_dispatcher/fuse_dispatcher.o 00:04:34.664 LIB libspdk_event.a 00:04:34.922 SO libspdk_event.so.14.0 00:04:34.922 SYMLINK libspdk_event.so 00:04:35.181 LIB libspdk_fuse_dispatcher.a 00:04:35.440 SO libspdk_fuse_dispatcher.so.1.0 00:04:35.440 SYMLINK libspdk_fuse_dispatcher.so 00:04:36.376 LIB libspdk_blob.a 00:04:36.376 SO libspdk_blob.so.12.0 00:04:36.635 SYMLINK libspdk_blob.so 00:04:36.895 CC lib/lvol/lvol.o 00:04:36.895 CC lib/blobfs/blobfs.o 00:04:36.895 CC lib/blobfs/tree.o 00:04:37.465 LIB libspdk_bdev.a 00:04:37.465 SO libspdk_bdev.so.17.0 00:04:37.465 SYMLINK libspdk_bdev.so 00:04:37.724 LIB libspdk_blobfs.a 00:04:37.724 CC lib/nbd/nbd.o 00:04:37.724 CC lib/scsi/lun.o 00:04:37.724 CC lib/scsi/port.o 00:04:37.724 CC lib/nbd/nbd_rpc.o 00:04:37.724 CC lib/scsi/dev.o 00:04:37.724 CC lib/ublk/ublk.o 00:04:37.724 CC lib/ftl/ftl_core.o 00:04:37.724 CC lib/nvmf/ctrlr.o 00:04:37.724 SO libspdk_blobfs.so.11.0 00:04:37.983 SYMLINK libspdk_blobfs.so 00:04:37.984 CC lib/nvmf/ctrlr_discovery.o 00:04:37.984 LIB libspdk_lvol.a 00:04:37.984 SO libspdk_lvol.so.11.0 00:04:37.984 CC lib/nvmf/ctrlr_bdev.o 00:04:37.984 CC lib/nvmf/subsystem.o 00:04:37.984 SYMLINK libspdk_lvol.so 00:04:37.984 CC lib/nvmf/nvmf.o 00:04:37.984 CC lib/nvmf/nvmf_rpc.o 00:04:37.984 CC lib/scsi/scsi.o 00:04:38.242 CC lib/ftl/ftl_init.o 00:04:38.242 LIB libspdk_nbd.a 00:04:38.242 CC lib/scsi/scsi_bdev.o 00:04:38.242 SO libspdk_nbd.so.7.0 00:04:38.242 SYMLINK libspdk_nbd.so 00:04:38.242 CC lib/scsi/scsi_pr.o 00:04:38.501 CC lib/ftl/ftl_layout.o 00:04:38.501 CC lib/nvmf/transport.o 00:04:38.501 CC lib/ublk/ublk_rpc.o 00:04:38.501 CC lib/ftl/ftl_debug.o 00:04:38.761 LIB libspdk_ublk.a 00:04:38.761 SO libspdk_ublk.so.3.0 00:04:38.761 CC lib/nvmf/tcp.o 00:04:38.761 CC lib/nvmf/stubs.o 00:04:38.761 SYMLINK libspdk_ublk.so 00:04:38.761 CC lib/ftl/ftl_io.o 00:04:38.761 CC lib/scsi/scsi_rpc.o 00:04:38.761 CC lib/ftl/ftl_sb.o 00:04:39.028 CC lib/scsi/task.o 00:04:39.028 CC lib/nvmf/mdns_server.o 00:04:39.028 CC lib/nvmf/rdma.o 00:04:39.028 CC lib/ftl/ftl_l2p.o 00:04:39.028 CC lib/ftl/ftl_l2p_flat.o 00:04:39.028 LIB libspdk_scsi.a 00:04:39.303 CC lib/nvmf/auth.o 00:04:39.303 SO libspdk_scsi.so.9.0 00:04:39.303 CC lib/ftl/ftl_nv_cache.o 00:04:39.303 CC lib/ftl/ftl_band.o 00:04:39.303 SYMLINK libspdk_scsi.so 00:04:39.303 CC lib/ftl/ftl_band_ops.o 00:04:39.303 CC lib/ftl/ftl_writer.o 00:04:39.303 CC lib/ftl/ftl_rq.o 00:04:39.563 CC lib/ftl/ftl_reloc.o 00:04:39.563 CC lib/ftl/ftl_l2p_cache.o 00:04:39.563 CC lib/ftl/ftl_p2l.o 00:04:39.563 CC lib/ftl/ftl_p2l_log.o 00:04:39.822 CC lib/ftl/mngt/ftl_mngt.o 00:04:39.822 CC lib/iscsi/conn.o 00:04:39.822 CC lib/ftl/mngt/ftl_mngt_bdev.o 00:04:40.082 CC lib/ftl/mngt/ftl_mngt_shutdown.o 00:04:40.082 CC lib/ftl/mngt/ftl_mngt_startup.o 00:04:40.082 CC lib/ftl/mngt/ftl_mngt_md.o 00:04:40.082 CC lib/ftl/mngt/ftl_mngt_misc.o 00:04:40.082 CC lib/iscsi/init_grp.o 00:04:40.082 CC lib/ftl/mngt/ftl_mngt_ioch.o 00:04:40.082 CC lib/ftl/mngt/ftl_mngt_l2p.o 00:04:40.342 CC lib/vhost/vhost.o 00:04:40.342 CC lib/ftl/mngt/ftl_mngt_band.o 00:04:40.342 CC lib/vhost/vhost_rpc.o 00:04:40.342 CC lib/vhost/vhost_scsi.o 00:04:40.342 CC lib/vhost/vhost_blk.o 00:04:40.342 CC lib/iscsi/iscsi.o 00:04:40.342 CC lib/iscsi/param.o 00:04:40.603 CC lib/vhost/rte_vhost_user.o 00:04:40.603 CC lib/ftl/mngt/ftl_mngt_self_test.o 00:04:40.603 CC lib/ftl/mngt/ftl_mngt_p2l.o 00:04:40.862 CC lib/iscsi/portal_grp.o 00:04:40.862 CC lib/iscsi/tgt_node.o 00:04:40.862 CC lib/ftl/mngt/ftl_mngt_recovery.o 00:04:41.122 CC lib/ftl/mngt/ftl_mngt_upgrade.o 00:04:41.122 CC lib/iscsi/iscsi_subsystem.o 00:04:41.122 CC lib/ftl/utils/ftl_conf.o 00:04:41.122 CC lib/iscsi/iscsi_rpc.o 00:04:41.382 CC lib/iscsi/task.o 00:04:41.382 CC lib/ftl/utils/ftl_md.o 00:04:41.382 CC lib/ftl/utils/ftl_mempool.o 00:04:41.382 CC lib/ftl/utils/ftl_bitmap.o 00:04:41.382 CC lib/ftl/utils/ftl_property.o 00:04:41.382 CC lib/ftl/utils/ftl_layout_tracker_bdev.o 00:04:41.641 CC lib/ftl/upgrade/ftl_layout_upgrade.o 00:04:41.641 CC lib/ftl/upgrade/ftl_sb_upgrade.o 00:04:41.641 LIB libspdk_nvmf.a 00:04:41.641 LIB libspdk_vhost.a 00:04:41.641 CC lib/ftl/upgrade/ftl_p2l_upgrade.o 00:04:41.641 SO libspdk_vhost.so.8.0 00:04:41.641 CC lib/ftl/upgrade/ftl_band_upgrade.o 00:04:41.641 CC lib/ftl/upgrade/ftl_chunk_upgrade.o 00:04:41.641 SO libspdk_nvmf.so.20.0 00:04:41.641 SYMLINK libspdk_vhost.so 00:04:41.641 CC lib/ftl/upgrade/ftl_trim_upgrade.o 00:04:41.641 CC lib/ftl/upgrade/ftl_sb_v3.o 00:04:41.641 CC lib/ftl/upgrade/ftl_sb_v5.o 00:04:41.901 CC lib/ftl/nvc/ftl_nvc_dev.o 00:04:41.901 CC lib/ftl/nvc/ftl_nvc_bdev_vss.o 00:04:41.901 CC lib/ftl/nvc/ftl_nvc_bdev_non_vss.o 00:04:41.901 CC lib/ftl/nvc/ftl_nvc_bdev_common.o 00:04:41.901 CC lib/ftl/base/ftl_base_dev.o 00:04:41.901 CC lib/ftl/base/ftl_base_bdev.o 00:04:41.901 SYMLINK libspdk_nvmf.so 00:04:41.901 CC lib/ftl/ftl_trace.o 00:04:42.160 LIB libspdk_iscsi.a 00:04:42.160 SO libspdk_iscsi.so.8.0 00:04:42.160 LIB libspdk_ftl.a 00:04:42.160 SYMLINK libspdk_iscsi.so 00:04:42.418 SO libspdk_ftl.so.9.0 00:04:42.676 SYMLINK libspdk_ftl.so 00:04:42.936 CC module/env_dpdk/env_dpdk_rpc.o 00:04:42.936 CC module/accel/error/accel_error.o 00:04:42.936 CC module/blob/bdev/blob_bdev.o 00:04:42.936 CC module/sock/posix/posix.o 00:04:42.936 CC module/scheduler/gscheduler/gscheduler.o 00:04:42.936 CC module/accel/ioat/accel_ioat.o 00:04:42.936 CC module/fsdev/aio/fsdev_aio.o 00:04:42.936 CC module/keyring/file/keyring.o 00:04:42.936 CC module/scheduler/dynamic/scheduler_dynamic.o 00:04:42.936 CC module/scheduler/dpdk_governor/dpdk_governor.o 00:04:43.196 LIB libspdk_env_dpdk_rpc.a 00:04:43.196 SO libspdk_env_dpdk_rpc.so.6.0 00:04:43.196 SYMLINK libspdk_env_dpdk_rpc.so 00:04:43.196 CC module/keyring/file/keyring_rpc.o 00:04:43.196 CC module/accel/error/accel_error_rpc.o 00:04:43.196 LIB libspdk_scheduler_gscheduler.a 00:04:43.196 LIB libspdk_scheduler_dpdk_governor.a 00:04:43.196 SO libspdk_scheduler_gscheduler.so.4.0 00:04:43.196 SO libspdk_scheduler_dpdk_governor.so.4.0 00:04:43.196 CC module/accel/ioat/accel_ioat_rpc.o 00:04:43.196 LIB libspdk_scheduler_dynamic.a 00:04:43.196 CC module/fsdev/aio/fsdev_aio_rpc.o 00:04:43.196 SYMLINK libspdk_scheduler_gscheduler.so 00:04:43.196 SO libspdk_scheduler_dynamic.so.4.0 00:04:43.196 SYMLINK libspdk_scheduler_dpdk_governor.so 00:04:43.196 LIB libspdk_blob_bdev.a 00:04:43.196 LIB libspdk_keyring_file.a 00:04:43.196 LIB libspdk_accel_error.a 00:04:43.196 SYMLINK libspdk_scheduler_dynamic.so 00:04:43.457 SO libspdk_blob_bdev.so.12.0 00:04:43.457 SO libspdk_keyring_file.so.2.0 00:04:43.457 LIB libspdk_accel_ioat.a 00:04:43.457 SO libspdk_accel_error.so.2.0 00:04:43.457 SO libspdk_accel_ioat.so.6.0 00:04:43.457 CC module/fsdev/aio/linux_aio_mgr.o 00:04:43.457 SYMLINK libspdk_blob_bdev.so 00:04:43.457 SYMLINK libspdk_keyring_file.so 00:04:43.457 SYMLINK libspdk_accel_error.so 00:04:43.457 CC module/keyring/linux/keyring.o 00:04:43.457 CC module/accel/dsa/accel_dsa.o 00:04:43.457 SYMLINK libspdk_accel_ioat.so 00:04:43.457 CC module/keyring/linux/keyring_rpc.o 00:04:43.457 CC module/accel/dsa/accel_dsa_rpc.o 00:04:43.457 CC module/accel/iaa/accel_iaa.o 00:04:43.717 LIB libspdk_keyring_linux.a 00:04:43.717 SO libspdk_keyring_linux.so.1.0 00:04:43.717 CC module/accel/iaa/accel_iaa_rpc.o 00:04:43.717 CC module/bdev/delay/vbdev_delay.o 00:04:43.717 CC module/blobfs/bdev/blobfs_bdev.o 00:04:43.717 SYMLINK libspdk_keyring_linux.so 00:04:43.717 CC module/blobfs/bdev/blobfs_bdev_rpc.o 00:04:43.717 CC module/bdev/error/vbdev_error.o 00:04:43.717 CC module/bdev/gpt/gpt.o 00:04:43.717 LIB libspdk_accel_dsa.a 00:04:43.717 LIB libspdk_accel_iaa.a 00:04:43.717 SO libspdk_accel_dsa.so.5.0 00:04:43.717 SO libspdk_accel_iaa.so.3.0 00:04:43.717 LIB libspdk_fsdev_aio.a 00:04:43.717 SO libspdk_fsdev_aio.so.1.0 00:04:43.717 SYMLINK libspdk_accel_dsa.so 00:04:43.717 LIB libspdk_sock_posix.a 00:04:43.977 CC module/bdev/error/vbdev_error_rpc.o 00:04:43.977 SYMLINK libspdk_accel_iaa.so 00:04:43.977 LIB libspdk_blobfs_bdev.a 00:04:43.977 CC module/bdev/lvol/vbdev_lvol.o 00:04:43.977 CC module/bdev/delay/vbdev_delay_rpc.o 00:04:43.977 CC module/bdev/lvol/vbdev_lvol_rpc.o 00:04:43.977 SO libspdk_sock_posix.so.6.0 00:04:43.977 SO libspdk_blobfs_bdev.so.6.0 00:04:43.977 SYMLINK libspdk_fsdev_aio.so 00:04:43.977 CC module/bdev/gpt/vbdev_gpt.o 00:04:43.977 SYMLINK libspdk_blobfs_bdev.so 00:04:43.977 SYMLINK libspdk_sock_posix.so 00:04:43.977 LIB libspdk_bdev_error.a 00:04:43.977 SO libspdk_bdev_error.so.6.0 00:04:43.977 LIB libspdk_bdev_delay.a 00:04:43.977 CC module/bdev/malloc/bdev_malloc.o 00:04:43.977 CC module/bdev/null/bdev_null.o 00:04:43.977 SO libspdk_bdev_delay.so.6.0 00:04:43.977 CC module/bdev/nvme/bdev_nvme.o 00:04:44.238 SYMLINK libspdk_bdev_error.so 00:04:44.238 CC module/bdev/nvme/bdev_nvme_rpc.o 00:04:44.238 CC module/bdev/passthru/vbdev_passthru.o 00:04:44.238 SYMLINK libspdk_bdev_delay.so 00:04:44.238 CC module/bdev/nvme/nvme_rpc.o 00:04:44.238 CC module/bdev/raid/bdev_raid.o 00:04:44.238 LIB libspdk_bdev_gpt.a 00:04:44.238 SO libspdk_bdev_gpt.so.6.0 00:04:44.238 CC module/bdev/null/bdev_null_rpc.o 00:04:44.238 SYMLINK libspdk_bdev_gpt.so 00:04:44.238 CC module/bdev/raid/bdev_raid_rpc.o 00:04:44.498 CC module/bdev/raid/bdev_raid_sb.o 00:04:44.498 CC module/bdev/raid/raid0.o 00:04:44.498 LIB libspdk_bdev_lvol.a 00:04:44.498 CC module/bdev/passthru/vbdev_passthru_rpc.o 00:04:44.498 CC module/bdev/malloc/bdev_malloc_rpc.o 00:04:44.498 SO libspdk_bdev_lvol.so.6.0 00:04:44.498 LIB libspdk_bdev_null.a 00:04:44.498 SO libspdk_bdev_null.so.6.0 00:04:44.498 SYMLINK libspdk_bdev_lvol.so 00:04:44.498 SYMLINK libspdk_bdev_null.so 00:04:44.498 CC module/bdev/raid/raid1.o 00:04:44.498 LIB libspdk_bdev_passthru.a 00:04:44.758 LIB libspdk_bdev_malloc.a 00:04:44.758 SO libspdk_bdev_passthru.so.6.0 00:04:44.758 SO libspdk_bdev_malloc.so.6.0 00:04:44.758 CC module/bdev/raid/concat.o 00:04:44.758 CC module/bdev/nvme/bdev_mdns_client.o 00:04:44.758 CC module/bdev/zone_block/vbdev_zone_block.o 00:04:44.758 SYMLINK libspdk_bdev_passthru.so 00:04:44.758 CC module/bdev/zone_block/vbdev_zone_block_rpc.o 00:04:44.758 CC module/bdev/split/vbdev_split.o 00:04:44.758 SYMLINK libspdk_bdev_malloc.so 00:04:44.758 CC module/bdev/nvme/vbdev_opal.o 00:04:44.758 CC module/bdev/aio/bdev_aio.o 00:04:44.758 CC module/bdev/aio/bdev_aio_rpc.o 00:04:45.017 CC module/bdev/raid/raid5f.o 00:04:45.017 CC module/bdev/ftl/bdev_ftl.o 00:04:45.017 CC module/bdev/split/vbdev_split_rpc.o 00:04:45.017 LIB libspdk_bdev_zone_block.a 00:04:45.017 CC module/bdev/iscsi/bdev_iscsi.o 00:04:45.017 SO libspdk_bdev_zone_block.so.6.0 00:04:45.017 LIB libspdk_bdev_split.a 00:04:45.017 CC module/bdev/nvme/vbdev_opal_rpc.o 00:04:45.017 SO libspdk_bdev_split.so.6.0 00:04:45.017 SYMLINK libspdk_bdev_zone_block.so 00:04:45.276 CC module/bdev/iscsi/bdev_iscsi_rpc.o 00:04:45.276 CC module/bdev/virtio/bdev_virtio_scsi.o 00:04:45.276 SYMLINK libspdk_bdev_split.so 00:04:45.276 CC module/bdev/ftl/bdev_ftl_rpc.o 00:04:45.276 LIB libspdk_bdev_aio.a 00:04:45.276 CC module/bdev/nvme/bdev_nvme_cuse_rpc.o 00:04:45.276 SO libspdk_bdev_aio.so.6.0 00:04:45.276 CC module/bdev/virtio/bdev_virtio_blk.o 00:04:45.276 SYMLINK libspdk_bdev_aio.so 00:04:45.276 CC module/bdev/virtio/bdev_virtio_rpc.o 00:04:45.276 LIB libspdk_bdev_ftl.a 00:04:45.536 LIB libspdk_bdev_raid.a 00:04:45.536 SO libspdk_bdev_ftl.so.6.0 00:04:45.536 LIB libspdk_bdev_iscsi.a 00:04:45.536 SO libspdk_bdev_iscsi.so.6.0 00:04:45.536 SYMLINK libspdk_bdev_ftl.so 00:04:45.536 SO libspdk_bdev_raid.so.6.0 00:04:45.536 SYMLINK libspdk_bdev_iscsi.so 00:04:45.536 SYMLINK libspdk_bdev_raid.so 00:04:45.796 LIB libspdk_bdev_virtio.a 00:04:45.796 SO libspdk_bdev_virtio.so.6.0 00:04:45.796 SYMLINK libspdk_bdev_virtio.so 00:04:46.735 LIB libspdk_bdev_nvme.a 00:04:46.996 SO libspdk_bdev_nvme.so.7.1 00:04:46.996 SYMLINK libspdk_bdev_nvme.so 00:04:47.565 CC module/event/subsystems/sock/sock.o 00:04:47.565 CC module/event/subsystems/fsdev/fsdev.o 00:04:47.565 CC module/event/subsystems/iobuf/iobuf_rpc.o 00:04:47.565 CC module/event/subsystems/iobuf/iobuf.o 00:04:47.565 CC module/event/subsystems/vmd/vmd.o 00:04:47.565 CC module/event/subsystems/vhost_blk/vhost_blk.o 00:04:47.565 CC module/event/subsystems/vmd/vmd_rpc.o 00:04:47.565 CC module/event/subsystems/keyring/keyring.o 00:04:47.565 CC module/event/subsystems/scheduler/scheduler.o 00:04:47.824 LIB libspdk_event_keyring.a 00:04:47.824 LIB libspdk_event_sock.a 00:04:47.824 LIB libspdk_event_vhost_blk.a 00:04:47.824 LIB libspdk_event_vmd.a 00:04:47.824 LIB libspdk_event_fsdev.a 00:04:47.824 SO libspdk_event_sock.so.5.0 00:04:47.824 SO libspdk_event_keyring.so.1.0 00:04:47.824 SO libspdk_event_vhost_blk.so.3.0 00:04:47.824 LIB libspdk_event_scheduler.a 00:04:47.824 SO libspdk_event_vmd.so.6.0 00:04:47.824 LIB libspdk_event_iobuf.a 00:04:47.824 SO libspdk_event_fsdev.so.1.0 00:04:47.824 SO libspdk_event_scheduler.so.4.0 00:04:47.824 SO libspdk_event_iobuf.so.3.0 00:04:47.824 SYMLINK libspdk_event_sock.so 00:04:47.824 SYMLINK libspdk_event_vhost_blk.so 00:04:47.824 SYMLINK libspdk_event_keyring.so 00:04:47.824 SYMLINK libspdk_event_vmd.so 00:04:47.824 SYMLINK libspdk_event_fsdev.so 00:04:47.824 SYMLINK libspdk_event_scheduler.so 00:04:47.824 SYMLINK libspdk_event_iobuf.so 00:04:48.084 CC module/event/subsystems/accel/accel.o 00:04:48.344 LIB libspdk_event_accel.a 00:04:48.344 SO libspdk_event_accel.so.6.0 00:04:48.344 SYMLINK libspdk_event_accel.so 00:04:48.913 CC module/event/subsystems/bdev/bdev.o 00:04:49.172 LIB libspdk_event_bdev.a 00:04:49.172 SO libspdk_event_bdev.so.6.0 00:04:49.172 SYMLINK libspdk_event_bdev.so 00:04:49.432 CC module/event/subsystems/nbd/nbd.o 00:04:49.432 CC module/event/subsystems/scsi/scsi.o 00:04:49.432 CC module/event/subsystems/ublk/ublk.o 00:04:49.432 CC module/event/subsystems/nvmf/nvmf_rpc.o 00:04:49.432 CC module/event/subsystems/nvmf/nvmf_tgt.o 00:04:49.693 LIB libspdk_event_nbd.a 00:04:49.693 SO libspdk_event_nbd.so.6.0 00:04:49.693 LIB libspdk_event_ublk.a 00:04:49.693 LIB libspdk_event_scsi.a 00:04:49.693 SO libspdk_event_ublk.so.3.0 00:04:49.693 SYMLINK libspdk_event_nbd.so 00:04:49.693 SO libspdk_event_scsi.so.6.0 00:04:49.693 SYMLINK libspdk_event_ublk.so 00:04:49.693 SYMLINK libspdk_event_scsi.so 00:04:49.693 LIB libspdk_event_nvmf.a 00:04:49.693 SO libspdk_event_nvmf.so.6.0 00:04:49.954 SYMLINK libspdk_event_nvmf.so 00:04:50.213 CC module/event/subsystems/iscsi/iscsi.o 00:04:50.214 CC module/event/subsystems/vhost_scsi/vhost_scsi.o 00:04:50.214 LIB libspdk_event_vhost_scsi.a 00:04:50.214 LIB libspdk_event_iscsi.a 00:04:50.214 SO libspdk_event_vhost_scsi.so.3.0 00:04:50.214 SO libspdk_event_iscsi.so.6.0 00:04:50.474 SYMLINK libspdk_event_vhost_scsi.so 00:04:50.474 SYMLINK libspdk_event_iscsi.so 00:04:50.474 SO libspdk.so.6.0 00:04:50.474 SYMLINK libspdk.so 00:04:51.042 CXX app/trace/trace.o 00:04:51.042 CC app/trace_record/trace_record.o 00:04:51.042 CC examples/interrupt_tgt/interrupt_tgt.o 00:04:51.042 CC app/iscsi_tgt/iscsi_tgt.o 00:04:51.042 CC app/nvmf_tgt/nvmf_main.o 00:04:51.042 CC examples/util/zipf/zipf.o 00:04:51.042 CC examples/ioat/perf/perf.o 00:04:51.042 CC test/thread/poller_perf/poller_perf.o 00:04:51.042 CC test/app/bdev_svc/bdev_svc.o 00:04:51.042 CC test/dma/test_dma/test_dma.o 00:04:51.042 LINK interrupt_tgt 00:04:51.042 LINK poller_perf 00:04:51.042 LINK zipf 00:04:51.042 LINK nvmf_tgt 00:04:51.042 LINK iscsi_tgt 00:04:51.042 LINK ioat_perf 00:04:51.042 LINK spdk_trace_record 00:04:51.301 LINK bdev_svc 00:04:51.301 LINK spdk_trace 00:04:51.301 CC examples/ioat/verify/verify.o 00:04:51.301 CC app/spdk_lspci/spdk_lspci.o 00:04:51.301 CC test/app/histogram_perf/histogram_perf.o 00:04:51.301 CC app/spdk_nvme_perf/perf.o 00:04:51.301 CC app/spdk_tgt/spdk_tgt.o 00:04:51.558 CC test/app/fuzz/iscsi_fuzz/iscsi_fuzz.o 00:04:51.558 CC examples/thread/thread/thread_ex.o 00:04:51.558 CC test/app/fuzz/nvme_fuzz/nvme_fuzz.o 00:04:51.558 CC test/app/fuzz/vhost_fuzz/vhost_fuzz_rpc.o 00:04:51.558 LINK spdk_lspci 00:04:51.558 LINK verify 00:04:51.558 LINK histogram_perf 00:04:51.558 LINK test_dma 00:04:51.558 LINK spdk_tgt 00:04:51.558 CC test/app/fuzz/vhost_fuzz/vhost_fuzz.o 00:04:51.817 LINK thread 00:04:51.817 CC app/spdk_nvme_identify/identify.o 00:04:51.817 CC app/spdk_nvme_discover/discovery_aer.o 00:04:51.817 CC examples/sock/hello_world/hello_sock.o 00:04:51.817 CC app/spdk_top/spdk_top.o 00:04:51.817 LINK nvme_fuzz 00:04:52.077 CC examples/vmd/lsvmd/lsvmd.o 00:04:52.077 LINK spdk_nvme_discover 00:04:52.077 CC examples/vmd/led/led.o 00:04:52.077 LINK lsvmd 00:04:52.077 LINK vhost_fuzz 00:04:52.077 LINK hello_sock 00:04:52.077 LINK led 00:04:52.337 CC app/vhost/vhost.o 00:04:52.337 CC test/app/jsoncat/jsoncat.o 00:04:52.337 LINK spdk_nvme_perf 00:04:52.337 CC test/app/stub/stub.o 00:04:52.337 TEST_HEADER include/spdk/accel.h 00:04:52.337 TEST_HEADER include/spdk/accel_module.h 00:04:52.337 TEST_HEADER include/spdk/assert.h 00:04:52.337 TEST_HEADER include/spdk/barrier.h 00:04:52.337 TEST_HEADER include/spdk/base64.h 00:04:52.337 TEST_HEADER include/spdk/bdev.h 00:04:52.337 TEST_HEADER include/spdk/bdev_module.h 00:04:52.337 TEST_HEADER include/spdk/bdev_zone.h 00:04:52.337 TEST_HEADER include/spdk/bit_array.h 00:04:52.337 TEST_HEADER include/spdk/bit_pool.h 00:04:52.337 TEST_HEADER include/spdk/blob_bdev.h 00:04:52.337 TEST_HEADER include/spdk/blobfs_bdev.h 00:04:52.337 TEST_HEADER include/spdk/blobfs.h 00:04:52.337 TEST_HEADER include/spdk/blob.h 00:04:52.337 TEST_HEADER include/spdk/conf.h 00:04:52.337 LINK jsoncat 00:04:52.337 TEST_HEADER include/spdk/config.h 00:04:52.337 TEST_HEADER include/spdk/cpuset.h 00:04:52.337 TEST_HEADER include/spdk/crc16.h 00:04:52.337 TEST_HEADER include/spdk/crc32.h 00:04:52.337 TEST_HEADER include/spdk/crc64.h 00:04:52.337 TEST_HEADER include/spdk/dif.h 00:04:52.337 TEST_HEADER include/spdk/dma.h 00:04:52.337 TEST_HEADER include/spdk/endian.h 00:04:52.337 TEST_HEADER include/spdk/env_dpdk.h 00:04:52.337 TEST_HEADER include/spdk/env.h 00:04:52.337 TEST_HEADER include/spdk/event.h 00:04:52.337 TEST_HEADER include/spdk/fd_group.h 00:04:52.337 TEST_HEADER include/spdk/fd.h 00:04:52.337 TEST_HEADER include/spdk/file.h 00:04:52.337 TEST_HEADER include/spdk/fsdev.h 00:04:52.337 TEST_HEADER include/spdk/fsdev_module.h 00:04:52.337 TEST_HEADER include/spdk/ftl.h 00:04:52.337 TEST_HEADER include/spdk/fuse_dispatcher.h 00:04:52.337 TEST_HEADER include/spdk/gpt_spec.h 00:04:52.337 TEST_HEADER include/spdk/hexlify.h 00:04:52.337 TEST_HEADER include/spdk/histogram_data.h 00:04:52.337 TEST_HEADER include/spdk/idxd.h 00:04:52.337 TEST_HEADER include/spdk/idxd_spec.h 00:04:52.337 LINK vhost 00:04:52.337 TEST_HEADER include/spdk/init.h 00:04:52.337 TEST_HEADER include/spdk/ioat.h 00:04:52.337 TEST_HEADER include/spdk/ioat_spec.h 00:04:52.337 TEST_HEADER include/spdk/iscsi_spec.h 00:04:52.337 TEST_HEADER include/spdk/json.h 00:04:52.337 TEST_HEADER include/spdk/jsonrpc.h 00:04:52.337 TEST_HEADER include/spdk/keyring.h 00:04:52.337 TEST_HEADER include/spdk/keyring_module.h 00:04:52.337 TEST_HEADER include/spdk/likely.h 00:04:52.337 TEST_HEADER include/spdk/log.h 00:04:52.337 TEST_HEADER include/spdk/lvol.h 00:04:52.337 TEST_HEADER include/spdk/md5.h 00:04:52.337 TEST_HEADER include/spdk/memory.h 00:04:52.337 TEST_HEADER include/spdk/mmio.h 00:04:52.337 TEST_HEADER include/spdk/nbd.h 00:04:52.337 TEST_HEADER include/spdk/net.h 00:04:52.337 TEST_HEADER include/spdk/notify.h 00:04:52.337 TEST_HEADER include/spdk/nvme.h 00:04:52.337 TEST_HEADER include/spdk/nvme_intel.h 00:04:52.337 TEST_HEADER include/spdk/nvme_ocssd.h 00:04:52.337 TEST_HEADER include/spdk/nvme_ocssd_spec.h 00:04:52.337 TEST_HEADER include/spdk/nvme_spec.h 00:04:52.337 TEST_HEADER include/spdk/nvme_zns.h 00:04:52.337 TEST_HEADER include/spdk/nvmf_cmd.h 00:04:52.337 TEST_HEADER include/spdk/nvmf_fc_spec.h 00:04:52.337 TEST_HEADER include/spdk/nvmf.h 00:04:52.337 TEST_HEADER include/spdk/nvmf_spec.h 00:04:52.337 TEST_HEADER include/spdk/nvmf_transport.h 00:04:52.596 TEST_HEADER include/spdk/opal.h 00:04:52.596 TEST_HEADER include/spdk/opal_spec.h 00:04:52.596 TEST_HEADER include/spdk/pci_ids.h 00:04:52.596 TEST_HEADER include/spdk/pipe.h 00:04:52.596 TEST_HEADER include/spdk/queue.h 00:04:52.596 TEST_HEADER include/spdk/reduce.h 00:04:52.596 TEST_HEADER include/spdk/rpc.h 00:04:52.596 CC examples/idxd/perf/perf.o 00:04:52.596 TEST_HEADER include/spdk/scheduler.h 00:04:52.596 TEST_HEADER include/spdk/scsi.h 00:04:52.596 TEST_HEADER include/spdk/scsi_spec.h 00:04:52.596 TEST_HEADER include/spdk/sock.h 00:04:52.596 TEST_HEADER include/spdk/stdinc.h 00:04:52.596 TEST_HEADER include/spdk/string.h 00:04:52.596 TEST_HEADER include/spdk/thread.h 00:04:52.596 TEST_HEADER include/spdk/trace.h 00:04:52.596 TEST_HEADER include/spdk/trace_parser.h 00:04:52.596 TEST_HEADER include/spdk/tree.h 00:04:52.596 TEST_HEADER include/spdk/ublk.h 00:04:52.596 TEST_HEADER include/spdk/util.h 00:04:52.596 TEST_HEADER include/spdk/uuid.h 00:04:52.596 TEST_HEADER include/spdk/version.h 00:04:52.596 LINK stub 00:04:52.596 TEST_HEADER include/spdk/vfio_user_pci.h 00:04:52.596 TEST_HEADER include/spdk/vfio_user_spec.h 00:04:52.596 TEST_HEADER include/spdk/vhost.h 00:04:52.596 TEST_HEADER include/spdk/vmd.h 00:04:52.596 CC test/env/mem_callbacks/mem_callbacks.o 00:04:52.596 TEST_HEADER include/spdk/xor.h 00:04:52.596 TEST_HEADER include/spdk/zipf.h 00:04:52.596 CXX test/cpp_headers/accel.o 00:04:52.596 CC app/spdk_dd/spdk_dd.o 00:04:52.596 CC app/fio/nvme/fio_plugin.o 00:04:52.596 CXX test/cpp_headers/accel_module.o 00:04:52.596 LINK spdk_nvme_identify 00:04:52.855 CC app/fio/bdev/fio_plugin.o 00:04:52.855 CC test/env/vtophys/vtophys.o 00:04:52.855 LINK idxd_perf 00:04:52.855 CXX test/cpp_headers/assert.o 00:04:52.855 CXX test/cpp_headers/barrier.o 00:04:52.855 LINK vtophys 00:04:52.855 LINK spdk_top 00:04:53.114 LINK spdk_dd 00:04:53.114 LINK mem_callbacks 00:04:53.115 CXX test/cpp_headers/base64.o 00:04:53.115 CXX test/cpp_headers/bdev.o 00:04:53.115 CC test/env/env_dpdk_post_init/env_dpdk_post_init.o 00:04:53.115 CC test/env/memory/memory_ut.o 00:04:53.115 CC examples/fsdev/hello_world/hello_fsdev.o 00:04:53.374 CXX test/cpp_headers/bdev_module.o 00:04:53.374 LINK spdk_bdev 00:04:53.374 LINK env_dpdk_post_init 00:04:53.374 LINK spdk_nvme 00:04:53.374 CC test/event/event_perf/event_perf.o 00:04:53.374 LINK iscsi_fuzz 00:04:53.374 CC test/nvme/aer/aer.o 00:04:53.374 CC examples/accel/perf/accel_perf.o 00:04:53.374 CC test/nvme/reset/reset.o 00:04:53.374 CXX test/cpp_headers/bdev_zone.o 00:04:53.374 CC test/env/pci/pci_ut.o 00:04:53.374 LINK hello_fsdev 00:04:53.633 LINK event_perf 00:04:53.633 CC test/nvme/sgl/sgl.o 00:04:53.633 CXX test/cpp_headers/bit_array.o 00:04:53.633 CC test/nvme/e2edp/nvme_dp.o 00:04:53.633 LINK aer 00:04:53.633 LINK reset 00:04:53.633 CC test/event/reactor/reactor.o 00:04:53.892 CC test/event/reactor_perf/reactor_perf.o 00:04:53.892 CXX test/cpp_headers/bit_pool.o 00:04:53.892 LINK sgl 00:04:53.892 LINK reactor 00:04:53.892 LINK reactor_perf 00:04:53.892 LINK pci_ut 00:04:53.892 LINK accel_perf 00:04:53.892 LINK nvme_dp 00:04:53.892 CXX test/cpp_headers/blob_bdev.o 00:04:53.892 CC test/event/app_repeat/app_repeat.o 00:04:53.892 CC test/nvme/overhead/overhead.o 00:04:54.151 CC test/nvme/err_injection/err_injection.o 00:04:54.151 LINK app_repeat 00:04:54.151 CC test/nvme/startup/startup.o 00:04:54.151 CXX test/cpp_headers/blobfs_bdev.o 00:04:54.151 CC test/nvme/reserve/reserve.o 00:04:54.151 CC test/event/scheduler/scheduler.o 00:04:54.151 LINK err_injection 00:04:54.151 CC examples/blob/hello_world/hello_blob.o 00:04:54.151 LINK overhead 00:04:54.151 LINK startup 00:04:54.410 CXX test/cpp_headers/blobfs.o 00:04:54.410 CC examples/nvme/hello_world/hello_world.o 00:04:54.410 CC examples/nvme/reconnect/reconnect.o 00:04:54.410 LINK memory_ut 00:04:54.410 LINK reserve 00:04:54.410 LINK scheduler 00:04:54.410 CXX test/cpp_headers/blob.o 00:04:54.410 LINK hello_blob 00:04:54.410 CC examples/nvme/nvme_manage/nvme_manage.o 00:04:54.410 CC examples/nvme/arbitration/arbitration.o 00:04:54.669 LINK hello_world 00:04:54.669 CXX test/cpp_headers/conf.o 00:04:54.669 CC test/nvme/simple_copy/simple_copy.o 00:04:54.669 CC examples/bdev/hello_world/hello_bdev.o 00:04:54.669 CC examples/bdev/bdevperf/bdevperf.o 00:04:54.669 LINK reconnect 00:04:54.669 CC examples/nvme/hotplug/hotplug.o 00:04:54.669 CXX test/cpp_headers/config.o 00:04:54.669 CC examples/blob/cli/blobcli.o 00:04:54.669 CXX test/cpp_headers/cpuset.o 00:04:54.669 CC test/nvme/connect_stress/connect_stress.o 00:04:54.929 LINK arbitration 00:04:54.929 LINK hello_bdev 00:04:54.929 LINK simple_copy 00:04:54.929 CXX test/cpp_headers/crc16.o 00:04:54.929 CC examples/nvme/cmb_copy/cmb_copy.o 00:04:54.929 LINK connect_stress 00:04:54.929 LINK hotplug 00:04:54.929 LINK nvme_manage 00:04:54.929 CXX test/cpp_headers/crc32.o 00:04:55.220 LINK cmb_copy 00:04:55.220 CC test/nvme/boot_partition/boot_partition.o 00:04:55.220 CC test/nvme/compliance/nvme_compliance.o 00:04:55.220 CC examples/nvme/abort/abort.o 00:04:55.220 CXX test/cpp_headers/crc64.o 00:04:55.220 CC test/nvme/fused_ordering/fused_ordering.o 00:04:55.220 CC test/nvme/doorbell_aers/doorbell_aers.o 00:04:55.220 CC examples/nvme/pmr_persistence/pmr_persistence.o 00:04:55.220 LINK blobcli 00:04:55.220 LINK boot_partition 00:04:55.493 CXX test/cpp_headers/dif.o 00:04:55.493 CC test/nvme/fdp/fdp.o 00:04:55.493 LINK fused_ordering 00:04:55.493 LINK pmr_persistence 00:04:55.493 LINK doorbell_aers 00:04:55.493 CXX test/cpp_headers/dma.o 00:04:55.493 LINK nvme_compliance 00:04:55.493 LINK bdevperf 00:04:55.493 LINK abort 00:04:55.493 CC test/rpc_client/rpc_client_test.o 00:04:55.493 CC test/nvme/cuse/cuse.o 00:04:55.780 CXX test/cpp_headers/endian.o 00:04:55.780 CXX test/cpp_headers/env_dpdk.o 00:04:55.780 CXX test/cpp_headers/env.o 00:04:55.780 LINK fdp 00:04:55.780 LINK rpc_client_test 00:04:55.780 CC test/accel/dif/dif.o 00:04:55.780 CXX test/cpp_headers/event.o 00:04:55.780 CC test/blobfs/mkfs/mkfs.o 00:04:55.780 CXX test/cpp_headers/fd_group.o 00:04:55.780 CC test/lvol/esnap/esnap.o 00:04:55.780 CXX test/cpp_headers/fd.o 00:04:55.780 CXX test/cpp_headers/file.o 00:04:56.039 CXX test/cpp_headers/fsdev.o 00:04:56.039 CC examples/nvmf/nvmf/nvmf.o 00:04:56.039 CXX test/cpp_headers/fsdev_module.o 00:04:56.039 CXX test/cpp_headers/ftl.o 00:04:56.039 CXX test/cpp_headers/fuse_dispatcher.o 00:04:56.039 LINK mkfs 00:04:56.039 CXX test/cpp_headers/gpt_spec.o 00:04:56.039 CXX test/cpp_headers/hexlify.o 00:04:56.039 CXX test/cpp_headers/histogram_data.o 00:04:56.299 CXX test/cpp_headers/idxd.o 00:04:56.299 CXX test/cpp_headers/idxd_spec.o 00:04:56.299 CXX test/cpp_headers/init.o 00:04:56.299 LINK nvmf 00:04:56.299 CXX test/cpp_headers/ioat.o 00:04:56.299 CXX test/cpp_headers/ioat_spec.o 00:04:56.299 CXX test/cpp_headers/iscsi_spec.o 00:04:56.299 CXX test/cpp_headers/json.o 00:04:56.299 CXX test/cpp_headers/jsonrpc.o 00:04:56.299 CXX test/cpp_headers/keyring.o 00:04:56.558 CXX test/cpp_headers/keyring_module.o 00:04:56.558 CXX test/cpp_headers/likely.o 00:04:56.558 CXX test/cpp_headers/log.o 00:04:56.558 CXX test/cpp_headers/lvol.o 00:04:56.558 CXX test/cpp_headers/md5.o 00:04:56.558 CXX test/cpp_headers/memory.o 00:04:56.558 CXX test/cpp_headers/mmio.o 00:04:56.558 LINK dif 00:04:56.558 CXX test/cpp_headers/nbd.o 00:04:56.558 CXX test/cpp_headers/net.o 00:04:56.558 CXX test/cpp_headers/notify.o 00:04:56.558 CXX test/cpp_headers/nvme.o 00:04:56.558 CXX test/cpp_headers/nvme_intel.o 00:04:56.558 CXX test/cpp_headers/nvme_ocssd.o 00:04:56.818 CXX test/cpp_headers/nvme_ocssd_spec.o 00:04:56.818 CXX test/cpp_headers/nvme_spec.o 00:04:56.818 CXX test/cpp_headers/nvme_zns.o 00:04:56.818 CXX test/cpp_headers/nvmf_cmd.o 00:04:56.818 CXX test/cpp_headers/nvmf_fc_spec.o 00:04:56.818 CXX test/cpp_headers/nvmf.o 00:04:56.818 CXX test/cpp_headers/nvmf_spec.o 00:04:56.818 CXX test/cpp_headers/nvmf_transport.o 00:04:56.818 CXX test/cpp_headers/opal.o 00:04:56.818 CXX test/cpp_headers/opal_spec.o 00:04:56.818 LINK cuse 00:04:56.818 CXX test/cpp_headers/pci_ids.o 00:04:56.818 CXX test/cpp_headers/pipe.o 00:04:57.077 CC test/bdev/bdevio/bdevio.o 00:04:57.077 CXX test/cpp_headers/queue.o 00:04:57.077 CXX test/cpp_headers/reduce.o 00:04:57.077 CXX test/cpp_headers/rpc.o 00:04:57.077 CXX test/cpp_headers/scheduler.o 00:04:57.077 CXX test/cpp_headers/scsi.o 00:04:57.077 CXX test/cpp_headers/scsi_spec.o 00:04:57.077 CXX test/cpp_headers/sock.o 00:04:57.077 CXX test/cpp_headers/stdinc.o 00:04:57.077 CXX test/cpp_headers/string.o 00:04:57.077 CXX test/cpp_headers/thread.o 00:04:57.077 CXX test/cpp_headers/trace.o 00:04:57.077 CXX test/cpp_headers/trace_parser.o 00:04:57.336 CXX test/cpp_headers/tree.o 00:04:57.336 CXX test/cpp_headers/ublk.o 00:04:57.336 CXX test/cpp_headers/util.o 00:04:57.336 CXX test/cpp_headers/uuid.o 00:04:57.336 CXX test/cpp_headers/version.o 00:04:57.336 CXX test/cpp_headers/vfio_user_pci.o 00:04:57.336 CXX test/cpp_headers/vfio_user_spec.o 00:04:57.336 CXX test/cpp_headers/vhost.o 00:04:57.336 CXX test/cpp_headers/vmd.o 00:04:57.336 LINK bdevio 00:04:57.336 CXX test/cpp_headers/xor.o 00:04:57.336 CXX test/cpp_headers/zipf.o 00:05:01.534 LINK esnap 00:05:01.794 00:05:01.794 real 1m18.547s 00:05:01.794 user 6m7.512s 00:05:01.794 sys 1m6.951s 00:05:01.794 15:09:29 make -- common/autotest_common.sh@1130 -- $ xtrace_disable 00:05:01.794 15:09:29 make -- common/autotest_common.sh@10 -- $ set +x 00:05:01.794 ************************************ 00:05:01.794 END TEST make 00:05:01.794 ************************************ 00:05:01.794 15:09:29 -- spdk/autobuild.sh@1 -- $ stop_monitor_resources 00:05:01.794 15:09:29 -- pm/common@29 -- $ signal_monitor_resources TERM 00:05:01.794 15:09:29 -- pm/common@40 -- $ local monitor pid pids signal=TERM 00:05:01.794 15:09:29 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:05:01.794 15:09:29 -- pm/common@43 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/power/collect-cpu-load.pid ]] 00:05:01.794 15:09:29 -- pm/common@44 -- $ pid=6216 00:05:01.794 15:09:29 -- pm/common@50 -- $ kill -TERM 6216 00:05:01.794 15:09:29 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:05:01.794 15:09:29 -- pm/common@43 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/power/collect-vmstat.pid ]] 00:05:01.794 15:09:29 -- pm/common@44 -- $ pid=6218 00:05:01.794 15:09:29 -- pm/common@50 -- $ kill -TERM 6218 00:05:01.794 15:09:29 -- spdk/autorun.sh@26 -- $ (( SPDK_TEST_UNITTEST == 1 || SPDK_RUN_FUNCTIONAL_TEST == 1 )) 00:05:01.794 15:09:29 -- spdk/autorun.sh@27 -- $ sudo -E /home/vagrant/spdk_repo/spdk/autotest.sh /home/vagrant/spdk_repo/autorun-spdk.conf 00:05:01.794 15:09:29 -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:05:01.794 15:09:29 -- common/autotest_common.sh@1693 -- # lcov --version 00:05:01.794 15:09:29 -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:05:01.794 15:09:29 -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:05:01.794 15:09:29 -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:01.794 15:09:29 -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:01.794 15:09:29 -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:01.794 15:09:29 -- scripts/common.sh@336 -- # IFS=.-: 00:05:01.794 15:09:29 -- scripts/common.sh@336 -- # read -ra ver1 00:05:01.794 15:09:29 -- scripts/common.sh@337 -- # IFS=.-: 00:05:01.794 15:09:29 -- scripts/common.sh@337 -- # read -ra ver2 00:05:01.794 15:09:29 -- scripts/common.sh@338 -- # local 'op=<' 00:05:01.794 15:09:29 -- scripts/common.sh@340 -- # ver1_l=2 00:05:01.794 15:09:29 -- scripts/common.sh@341 -- # ver2_l=1 00:05:01.794 15:09:29 -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:02.055 15:09:29 -- scripts/common.sh@344 -- # case "$op" in 00:05:02.055 15:09:29 -- scripts/common.sh@345 -- # : 1 00:05:02.055 15:09:29 -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:02.055 15:09:29 -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:02.055 15:09:29 -- scripts/common.sh@365 -- # decimal 1 00:05:02.055 15:09:29 -- scripts/common.sh@353 -- # local d=1 00:05:02.055 15:09:29 -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:02.055 15:09:29 -- scripts/common.sh@355 -- # echo 1 00:05:02.055 15:09:29 -- scripts/common.sh@365 -- # ver1[v]=1 00:05:02.055 15:09:29 -- scripts/common.sh@366 -- # decimal 2 00:05:02.055 15:09:29 -- scripts/common.sh@353 -- # local d=2 00:05:02.055 15:09:29 -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:02.055 15:09:29 -- scripts/common.sh@355 -- # echo 2 00:05:02.055 15:09:29 -- scripts/common.sh@366 -- # ver2[v]=2 00:05:02.055 15:09:29 -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:02.055 15:09:29 -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:02.055 15:09:29 -- scripts/common.sh@368 -- # return 0 00:05:02.055 15:09:29 -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:02.055 15:09:29 -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:05:02.055 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:02.055 --rc genhtml_branch_coverage=1 00:05:02.055 --rc genhtml_function_coverage=1 00:05:02.055 --rc genhtml_legend=1 00:05:02.055 --rc geninfo_all_blocks=1 00:05:02.055 --rc geninfo_unexecuted_blocks=1 00:05:02.055 00:05:02.055 ' 00:05:02.055 15:09:29 -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:05:02.055 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:02.055 --rc genhtml_branch_coverage=1 00:05:02.055 --rc genhtml_function_coverage=1 00:05:02.055 --rc genhtml_legend=1 00:05:02.055 --rc geninfo_all_blocks=1 00:05:02.055 --rc geninfo_unexecuted_blocks=1 00:05:02.055 00:05:02.056 ' 00:05:02.056 15:09:29 -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:05:02.056 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:02.056 --rc genhtml_branch_coverage=1 00:05:02.056 --rc genhtml_function_coverage=1 00:05:02.056 --rc genhtml_legend=1 00:05:02.056 --rc geninfo_all_blocks=1 00:05:02.056 --rc geninfo_unexecuted_blocks=1 00:05:02.056 00:05:02.056 ' 00:05:02.056 15:09:29 -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:05:02.056 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:02.056 --rc genhtml_branch_coverage=1 00:05:02.056 --rc genhtml_function_coverage=1 00:05:02.056 --rc genhtml_legend=1 00:05:02.056 --rc geninfo_all_blocks=1 00:05:02.056 --rc geninfo_unexecuted_blocks=1 00:05:02.056 00:05:02.056 ' 00:05:02.056 15:09:29 -- spdk/autotest.sh@25 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:05:02.056 15:09:29 -- nvmf/common.sh@7 -- # uname -s 00:05:02.056 15:09:29 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:05:02.056 15:09:29 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:05:02.056 15:09:29 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:05:02.056 15:09:29 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:05:02.056 15:09:29 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:05:02.056 15:09:29 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:05:02.056 15:09:29 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:05:02.056 15:09:29 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:05:02.056 15:09:29 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:05:02.056 15:09:29 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:05:02.056 15:09:29 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:d42b44b2-5d93-4f0c-9e37-1c47f6668401 00:05:02.056 15:09:29 -- nvmf/common.sh@18 -- # NVME_HOSTID=d42b44b2-5d93-4f0c-9e37-1c47f6668401 00:05:02.056 15:09:29 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:05:02.056 15:09:29 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:05:02.056 15:09:29 -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:05:02.056 15:09:29 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:05:02.056 15:09:29 -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:05:02.056 15:09:29 -- scripts/common.sh@15 -- # shopt -s extglob 00:05:02.056 15:09:29 -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:05:02.056 15:09:29 -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:05:02.056 15:09:29 -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:05:02.056 15:09:29 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:02.056 15:09:29 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:02.056 15:09:29 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:02.056 15:09:29 -- paths/export.sh@5 -- # export PATH 00:05:02.056 15:09:29 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:02.056 15:09:29 -- nvmf/common.sh@51 -- # : 0 00:05:02.056 15:09:29 -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:05:02.056 15:09:29 -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:05:02.056 15:09:29 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:05:02.056 15:09:29 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:05:02.056 15:09:29 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:05:02.056 15:09:29 -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:05:02.056 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:05:02.056 15:09:29 -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:05:02.056 15:09:29 -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:05:02.056 15:09:29 -- nvmf/common.sh@55 -- # have_pci_nics=0 00:05:02.056 15:09:29 -- spdk/autotest.sh@27 -- # '[' 0 -ne 0 ']' 00:05:02.056 15:09:29 -- spdk/autotest.sh@32 -- # uname -s 00:05:02.056 15:09:29 -- spdk/autotest.sh@32 -- # '[' Linux = Linux ']' 00:05:02.056 15:09:29 -- spdk/autotest.sh@33 -- # old_core_pattern='|/usr/lib/systemd/systemd-coredump %P %u %g %s %t %c %h' 00:05:02.056 15:09:29 -- spdk/autotest.sh@34 -- # mkdir -p /home/vagrant/spdk_repo/spdk/../output/coredumps 00:05:02.056 15:09:29 -- spdk/autotest.sh@39 -- # echo '|/home/vagrant/spdk_repo/spdk/scripts/core-collector.sh %P %s %t' 00:05:02.056 15:09:29 -- spdk/autotest.sh@40 -- # echo /home/vagrant/spdk_repo/spdk/../output/coredumps 00:05:02.056 15:09:29 -- spdk/autotest.sh@44 -- # modprobe nbd 00:05:02.056 15:09:30 -- spdk/autotest.sh@46 -- # type -P udevadm 00:05:02.056 15:09:30 -- spdk/autotest.sh@46 -- # udevadm=/usr/sbin/udevadm 00:05:02.056 15:09:30 -- spdk/autotest.sh@48 -- # udevadm_pid=66911 00:05:02.056 15:09:30 -- spdk/autotest.sh@47 -- # /usr/sbin/udevadm monitor --property 00:05:02.056 15:09:30 -- spdk/autotest.sh@53 -- # start_monitor_resources 00:05:02.056 15:09:30 -- pm/common@17 -- # local monitor 00:05:02.056 15:09:30 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:05:02.056 15:09:30 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:05:02.056 15:09:30 -- pm/common@25 -- # sleep 1 00:05:02.056 15:09:30 -- pm/common@21 -- # date +%s 00:05:02.056 15:09:30 -- pm/common@21 -- # date +%s 00:05:02.056 15:09:30 -- pm/common@21 -- # /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-vmstat -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autotest.sh.1732720170 00:05:02.056 15:09:30 -- pm/common@21 -- # /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-cpu-load -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autotest.sh.1732720170 00:05:02.056 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autotest.sh.1732720170_collect-cpu-load.pm.log 00:05:02.056 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autotest.sh.1732720170_collect-vmstat.pm.log 00:05:02.994 15:09:31 -- spdk/autotest.sh@55 -- # trap 'autotest_cleanup || :; exit 1' SIGINT SIGTERM EXIT 00:05:02.994 15:09:31 -- spdk/autotest.sh@57 -- # timing_enter autotest 00:05:02.994 15:09:31 -- common/autotest_common.sh@726 -- # xtrace_disable 00:05:02.994 15:09:31 -- common/autotest_common.sh@10 -- # set +x 00:05:02.994 15:09:31 -- spdk/autotest.sh@59 -- # create_test_list 00:05:02.994 15:09:31 -- common/autotest_common.sh@752 -- # xtrace_disable 00:05:02.994 15:09:31 -- common/autotest_common.sh@10 -- # set +x 00:05:03.253 15:09:31 -- spdk/autotest.sh@61 -- # dirname /home/vagrant/spdk_repo/spdk/autotest.sh 00:05:03.253 15:09:31 -- spdk/autotest.sh@61 -- # readlink -f /home/vagrant/spdk_repo/spdk 00:05:03.253 15:09:31 -- spdk/autotest.sh@61 -- # src=/home/vagrant/spdk_repo/spdk 00:05:03.253 15:09:31 -- spdk/autotest.sh@62 -- # out=/home/vagrant/spdk_repo/spdk/../output 00:05:03.253 15:09:31 -- spdk/autotest.sh@63 -- # cd /home/vagrant/spdk_repo/spdk 00:05:03.253 15:09:31 -- spdk/autotest.sh@65 -- # freebsd_update_contigmem_mod 00:05:03.253 15:09:31 -- common/autotest_common.sh@1457 -- # uname 00:05:03.253 15:09:31 -- common/autotest_common.sh@1457 -- # '[' Linux = FreeBSD ']' 00:05:03.253 15:09:31 -- spdk/autotest.sh@66 -- # freebsd_set_maxsock_buf 00:05:03.253 15:09:31 -- common/autotest_common.sh@1477 -- # uname 00:05:03.253 15:09:31 -- common/autotest_common.sh@1477 -- # [[ Linux = FreeBSD ]] 00:05:03.253 15:09:31 -- spdk/autotest.sh@68 -- # [[ y == y ]] 00:05:03.253 15:09:31 -- spdk/autotest.sh@70 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 --version 00:05:03.253 lcov: LCOV version 1.15 00:05:03.253 15:09:31 -- spdk/autotest.sh@72 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -c --no-external -i -t Baseline -d /home/vagrant/spdk_repo/spdk -o /home/vagrant/spdk_repo/spdk/../output/cov_base.info 00:05:18.144 /home/vagrant/spdk_repo/spdk/lib/nvme/nvme_stubs.gcno:no functions found 00:05:18.144 geninfo: WARNING: GCOV did not produce any data for /home/vagrant/spdk_repo/spdk/lib/nvme/nvme_stubs.gcno 00:05:33.037 15:09:59 -- spdk/autotest.sh@76 -- # timing_enter pre_cleanup 00:05:33.037 15:09:59 -- common/autotest_common.sh@726 -- # xtrace_disable 00:05:33.037 15:09:59 -- common/autotest_common.sh@10 -- # set +x 00:05:33.037 15:09:59 -- spdk/autotest.sh@78 -- # rm -f 00:05:33.037 15:09:59 -- spdk/autotest.sh@81 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh reset 00:05:33.037 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:33.037 0000:00:11.0 (1b36 0010): Already using the nvme driver 00:05:33.037 0000:00:10.0 (1b36 0010): Already using the nvme driver 00:05:33.037 15:10:00 -- spdk/autotest.sh@83 -- # get_zoned_devs 00:05:33.037 15:10:00 -- common/autotest_common.sh@1657 -- # zoned_devs=() 00:05:33.037 15:10:00 -- common/autotest_common.sh@1657 -- # local -gA zoned_devs 00:05:33.037 15:10:00 -- common/autotest_common.sh@1658 -- # local nvme bdf 00:05:33.037 15:10:00 -- common/autotest_common.sh@1660 -- # for nvme in /sys/block/nvme* 00:05:33.037 15:10:00 -- common/autotest_common.sh@1661 -- # is_block_zoned nvme0n1 00:05:33.037 15:10:00 -- common/autotest_common.sh@1650 -- # local device=nvme0n1 00:05:33.037 15:10:00 -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:05:33.037 15:10:00 -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:05:33.037 15:10:00 -- common/autotest_common.sh@1660 -- # for nvme in /sys/block/nvme* 00:05:33.037 15:10:00 -- common/autotest_common.sh@1661 -- # is_block_zoned nvme1n1 00:05:33.037 15:10:00 -- common/autotest_common.sh@1650 -- # local device=nvme1n1 00:05:33.037 15:10:00 -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme1n1/queue/zoned ]] 00:05:33.037 15:10:00 -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:05:33.037 15:10:00 -- common/autotest_common.sh@1660 -- # for nvme in /sys/block/nvme* 00:05:33.038 15:10:00 -- common/autotest_common.sh@1661 -- # is_block_zoned nvme1n2 00:05:33.038 15:10:00 -- common/autotest_common.sh@1650 -- # local device=nvme1n2 00:05:33.038 15:10:00 -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme1n2/queue/zoned ]] 00:05:33.038 15:10:00 -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:05:33.038 15:10:00 -- common/autotest_common.sh@1660 -- # for nvme in /sys/block/nvme* 00:05:33.038 15:10:00 -- common/autotest_common.sh@1661 -- # is_block_zoned nvme1n3 00:05:33.038 15:10:00 -- common/autotest_common.sh@1650 -- # local device=nvme1n3 00:05:33.038 15:10:00 -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme1n3/queue/zoned ]] 00:05:33.038 15:10:00 -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:05:33.038 15:10:00 -- spdk/autotest.sh@85 -- # (( 0 > 0 )) 00:05:33.038 15:10:00 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:05:33.038 15:10:00 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:05:33.038 15:10:00 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme0n1 00:05:33.038 15:10:00 -- scripts/common.sh@381 -- # local block=/dev/nvme0n1 pt 00:05:33.038 15:10:00 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme0n1 00:05:33.038 No valid GPT data, bailing 00:05:33.038 15:10:00 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:05:33.038 15:10:00 -- scripts/common.sh@394 -- # pt= 00:05:33.038 15:10:00 -- scripts/common.sh@395 -- # return 1 00:05:33.038 15:10:00 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme0n1 bs=1M count=1 00:05:33.038 1+0 records in 00:05:33.038 1+0 records out 00:05:33.038 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00489202 s, 214 MB/s 00:05:33.038 15:10:00 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:05:33.038 15:10:00 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:05:33.038 15:10:00 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme1n1 00:05:33.038 15:10:00 -- scripts/common.sh@381 -- # local block=/dev/nvme1n1 pt 00:05:33.038 15:10:00 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme1n1 00:05:33.038 No valid GPT data, bailing 00:05:33.038 15:10:00 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme1n1 00:05:33.038 15:10:00 -- scripts/common.sh@394 -- # pt= 00:05:33.038 15:10:00 -- scripts/common.sh@395 -- # return 1 00:05:33.038 15:10:00 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme1n1 bs=1M count=1 00:05:33.038 1+0 records in 00:05:33.038 1+0 records out 00:05:33.038 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00508101 s, 206 MB/s 00:05:33.038 15:10:00 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:05:33.038 15:10:00 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:05:33.038 15:10:00 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme1n2 00:05:33.038 15:10:00 -- scripts/common.sh@381 -- # local block=/dev/nvme1n2 pt 00:05:33.038 15:10:00 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme1n2 00:05:33.038 No valid GPT data, bailing 00:05:33.038 15:10:01 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme1n2 00:05:33.038 15:10:01 -- scripts/common.sh@394 -- # pt= 00:05:33.038 15:10:01 -- scripts/common.sh@395 -- # return 1 00:05:33.038 15:10:01 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme1n2 bs=1M count=1 00:05:33.038 1+0 records in 00:05:33.038 1+0 records out 00:05:33.038 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00605672 s, 173 MB/s 00:05:33.038 15:10:01 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:05:33.038 15:10:01 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:05:33.038 15:10:01 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme1n3 00:05:33.038 15:10:01 -- scripts/common.sh@381 -- # local block=/dev/nvme1n3 pt 00:05:33.038 15:10:01 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme1n3 00:05:33.038 No valid GPT data, bailing 00:05:33.038 15:10:01 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme1n3 00:05:33.038 15:10:01 -- scripts/common.sh@394 -- # pt= 00:05:33.038 15:10:01 -- scripts/common.sh@395 -- # return 1 00:05:33.038 15:10:01 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme1n3 bs=1M count=1 00:05:33.038 1+0 records in 00:05:33.038 1+0 records out 00:05:33.038 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00407203 s, 258 MB/s 00:05:33.038 15:10:01 -- spdk/autotest.sh@105 -- # sync 00:05:33.297 15:10:01 -- spdk/autotest.sh@107 -- # xtrace_disable_per_cmd reap_spdk_processes 00:05:33.297 15:10:01 -- common/autotest_common.sh@22 -- # eval 'reap_spdk_processes 12> /dev/null' 00:05:33.297 15:10:01 -- common/autotest_common.sh@22 -- # reap_spdk_processes 00:05:36.616 15:10:04 -- spdk/autotest.sh@111 -- # uname -s 00:05:36.616 15:10:04 -- spdk/autotest.sh@111 -- # [[ Linux == Linux ]] 00:05:36.616 15:10:04 -- spdk/autotest.sh@111 -- # [[ 0 -eq 1 ]] 00:05:36.616 15:10:04 -- spdk/autotest.sh@115 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh status 00:05:37.187 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:37.187 Hugepages 00:05:37.187 node hugesize free / total 00:05:37.187 node0 1048576kB 0 / 0 00:05:37.187 node0 2048kB 0 / 0 00:05:37.187 00:05:37.187 Type BDF Vendor Device NUMA Driver Device Block devices 00:05:37.187 virtio 0000:00:03.0 1af4 1001 unknown virtio-pci - vda 00:05:37.446 NVMe 0000:00:10.0 1b36 0010 unknown nvme nvme0 nvme0n1 00:05:37.446 NVMe 0000:00:11.0 1b36 0010 unknown nvme nvme1 nvme1n1 nvme1n2 nvme1n3 00:05:37.446 15:10:05 -- spdk/autotest.sh@117 -- # uname -s 00:05:37.446 15:10:05 -- spdk/autotest.sh@117 -- # [[ Linux == Linux ]] 00:05:37.446 15:10:05 -- spdk/autotest.sh@119 -- # nvme_namespace_revert 00:05:37.446 15:10:05 -- common/autotest_common.sh@1516 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh 00:05:38.384 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:38.384 0000:00:10.0 (1b36 0010): nvme -> uio_pci_generic 00:05:38.384 0000:00:11.0 (1b36 0010): nvme -> uio_pci_generic 00:05:38.644 15:10:06 -- common/autotest_common.sh@1517 -- # sleep 1 00:05:39.582 15:10:07 -- common/autotest_common.sh@1518 -- # bdfs=() 00:05:39.582 15:10:07 -- common/autotest_common.sh@1518 -- # local bdfs 00:05:39.582 15:10:07 -- common/autotest_common.sh@1520 -- # bdfs=($(get_nvme_bdfs)) 00:05:39.582 15:10:07 -- common/autotest_common.sh@1520 -- # get_nvme_bdfs 00:05:39.582 15:10:07 -- common/autotest_common.sh@1498 -- # bdfs=() 00:05:39.582 15:10:07 -- common/autotest_common.sh@1498 -- # local bdfs 00:05:39.582 15:10:07 -- common/autotest_common.sh@1499 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:05:39.582 15:10:07 -- common/autotest_common.sh@1499 -- # /home/vagrant/spdk_repo/spdk/scripts/gen_nvme.sh 00:05:39.582 15:10:07 -- common/autotest_common.sh@1499 -- # jq -r '.config[].params.traddr' 00:05:39.582 15:10:07 -- common/autotest_common.sh@1500 -- # (( 2 == 0 )) 00:05:39.582 15:10:07 -- common/autotest_common.sh@1504 -- # printf '%s\n' 0000:00:10.0 0000:00:11.0 00:05:39.582 15:10:07 -- common/autotest_common.sh@1522 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh reset 00:05:40.151 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:40.151 Waiting for block devices as requested 00:05:40.151 0000:00:11.0 (1b36 0010): uio_pci_generic -> nvme 00:05:40.412 0000:00:10.0 (1b36 0010): uio_pci_generic -> nvme 00:05:40.412 15:10:08 -- common/autotest_common.sh@1524 -- # for bdf in "${bdfs[@]}" 00:05:40.412 15:10:08 -- common/autotest_common.sh@1525 -- # get_nvme_ctrlr_from_bdf 0000:00:10.0 00:05:40.412 15:10:08 -- common/autotest_common.sh@1487 -- # readlink -f /sys/class/nvme/nvme0 /sys/class/nvme/nvme1 00:05:40.412 15:10:08 -- common/autotest_common.sh@1487 -- # grep 0000:00:10.0/nvme/nvme 00:05:40.412 15:10:08 -- common/autotest_common.sh@1487 -- # bdf_sysfs_path=/sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 00:05:40.412 15:10:08 -- common/autotest_common.sh@1488 -- # [[ -z /sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 ]] 00:05:40.412 15:10:08 -- common/autotest_common.sh@1492 -- # basename /sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 00:05:40.412 15:10:08 -- common/autotest_common.sh@1492 -- # printf '%s\n' nvme1 00:05:40.412 15:10:08 -- common/autotest_common.sh@1525 -- # nvme_ctrlr=/dev/nvme1 00:05:40.412 15:10:08 -- common/autotest_common.sh@1526 -- # [[ -z /dev/nvme1 ]] 00:05:40.412 15:10:08 -- common/autotest_common.sh@1531 -- # nvme id-ctrl /dev/nvme1 00:05:40.412 15:10:08 -- common/autotest_common.sh@1531 -- # grep oacs 00:05:40.412 15:10:08 -- common/autotest_common.sh@1531 -- # cut -d: -f2 00:05:40.412 15:10:08 -- common/autotest_common.sh@1531 -- # oacs=' 0x12a' 00:05:40.412 15:10:08 -- common/autotest_common.sh@1532 -- # oacs_ns_manage=8 00:05:40.412 15:10:08 -- common/autotest_common.sh@1534 -- # [[ 8 -ne 0 ]] 00:05:40.412 15:10:08 -- common/autotest_common.sh@1540 -- # nvme id-ctrl /dev/nvme1 00:05:40.412 15:10:08 -- common/autotest_common.sh@1540 -- # grep unvmcap 00:05:40.412 15:10:08 -- common/autotest_common.sh@1540 -- # cut -d: -f2 00:05:40.412 15:10:08 -- common/autotest_common.sh@1540 -- # unvmcap=' 0' 00:05:40.412 15:10:08 -- common/autotest_common.sh@1541 -- # [[ 0 -eq 0 ]] 00:05:40.412 15:10:08 -- common/autotest_common.sh@1543 -- # continue 00:05:40.412 15:10:08 -- common/autotest_common.sh@1524 -- # for bdf in "${bdfs[@]}" 00:05:40.412 15:10:08 -- common/autotest_common.sh@1525 -- # get_nvme_ctrlr_from_bdf 0000:00:11.0 00:05:40.412 15:10:08 -- common/autotest_common.sh@1487 -- # readlink -f /sys/class/nvme/nvme0 /sys/class/nvme/nvme1 00:05:40.412 15:10:08 -- common/autotest_common.sh@1487 -- # grep 0000:00:11.0/nvme/nvme 00:05:40.412 15:10:08 -- common/autotest_common.sh@1487 -- # bdf_sysfs_path=/sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 00:05:40.412 15:10:08 -- common/autotest_common.sh@1488 -- # [[ -z /sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 ]] 00:05:40.412 15:10:08 -- common/autotest_common.sh@1492 -- # basename /sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 00:05:40.412 15:10:08 -- common/autotest_common.sh@1492 -- # printf '%s\n' nvme0 00:05:40.412 15:10:08 -- common/autotest_common.sh@1525 -- # nvme_ctrlr=/dev/nvme0 00:05:40.412 15:10:08 -- common/autotest_common.sh@1526 -- # [[ -z /dev/nvme0 ]] 00:05:40.412 15:10:08 -- common/autotest_common.sh@1531 -- # nvme id-ctrl /dev/nvme0 00:05:40.412 15:10:08 -- common/autotest_common.sh@1531 -- # grep oacs 00:05:40.412 15:10:08 -- common/autotest_common.sh@1531 -- # cut -d: -f2 00:05:40.412 15:10:08 -- common/autotest_common.sh@1531 -- # oacs=' 0x12a' 00:05:40.412 15:10:08 -- common/autotest_common.sh@1532 -- # oacs_ns_manage=8 00:05:40.412 15:10:08 -- common/autotest_common.sh@1534 -- # [[ 8 -ne 0 ]] 00:05:40.412 15:10:08 -- common/autotest_common.sh@1540 -- # nvme id-ctrl /dev/nvme0 00:05:40.412 15:10:08 -- common/autotest_common.sh@1540 -- # grep unvmcap 00:05:40.412 15:10:08 -- common/autotest_common.sh@1540 -- # cut -d: -f2 00:05:40.412 15:10:08 -- common/autotest_common.sh@1540 -- # unvmcap=' 0' 00:05:40.412 15:10:08 -- common/autotest_common.sh@1541 -- # [[ 0 -eq 0 ]] 00:05:40.412 15:10:08 -- common/autotest_common.sh@1543 -- # continue 00:05:40.412 15:10:08 -- spdk/autotest.sh@122 -- # timing_exit pre_cleanup 00:05:40.412 15:10:08 -- common/autotest_common.sh@732 -- # xtrace_disable 00:05:40.412 15:10:08 -- common/autotest_common.sh@10 -- # set +x 00:05:40.672 15:10:08 -- spdk/autotest.sh@125 -- # timing_enter afterboot 00:05:40.672 15:10:08 -- common/autotest_common.sh@726 -- # xtrace_disable 00:05:40.672 15:10:08 -- common/autotest_common.sh@10 -- # set +x 00:05:40.672 15:10:08 -- spdk/autotest.sh@126 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh 00:05:41.251 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:41.512 0000:00:10.0 (1b36 0010): nvme -> uio_pci_generic 00:05:41.512 0000:00:11.0 (1b36 0010): nvme -> uio_pci_generic 00:05:41.512 15:10:09 -- spdk/autotest.sh@127 -- # timing_exit afterboot 00:05:41.512 15:10:09 -- common/autotest_common.sh@732 -- # xtrace_disable 00:05:41.512 15:10:09 -- common/autotest_common.sh@10 -- # set +x 00:05:41.772 15:10:09 -- spdk/autotest.sh@131 -- # opal_revert_cleanup 00:05:41.773 15:10:09 -- common/autotest_common.sh@1578 -- # mapfile -t bdfs 00:05:41.773 15:10:09 -- common/autotest_common.sh@1578 -- # get_nvme_bdfs_by_id 0x0a54 00:05:41.773 15:10:09 -- common/autotest_common.sh@1563 -- # bdfs=() 00:05:41.773 15:10:09 -- common/autotest_common.sh@1563 -- # _bdfs=() 00:05:41.773 15:10:09 -- common/autotest_common.sh@1563 -- # local bdfs _bdfs 00:05:41.773 15:10:09 -- common/autotest_common.sh@1564 -- # _bdfs=($(get_nvme_bdfs)) 00:05:41.773 15:10:09 -- common/autotest_common.sh@1564 -- # get_nvme_bdfs 00:05:41.773 15:10:09 -- common/autotest_common.sh@1498 -- # bdfs=() 00:05:41.773 15:10:09 -- common/autotest_common.sh@1498 -- # local bdfs 00:05:41.773 15:10:09 -- common/autotest_common.sh@1499 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:05:41.773 15:10:09 -- common/autotest_common.sh@1499 -- # /home/vagrant/spdk_repo/spdk/scripts/gen_nvme.sh 00:05:41.773 15:10:09 -- common/autotest_common.sh@1499 -- # jq -r '.config[].params.traddr' 00:05:41.773 15:10:09 -- common/autotest_common.sh@1500 -- # (( 2 == 0 )) 00:05:41.773 15:10:09 -- common/autotest_common.sh@1504 -- # printf '%s\n' 0000:00:10.0 0000:00:11.0 00:05:41.773 15:10:09 -- common/autotest_common.sh@1565 -- # for bdf in "${_bdfs[@]}" 00:05:41.773 15:10:09 -- common/autotest_common.sh@1566 -- # cat /sys/bus/pci/devices/0000:00:10.0/device 00:05:41.773 15:10:09 -- common/autotest_common.sh@1566 -- # device=0x0010 00:05:41.773 15:10:09 -- common/autotest_common.sh@1567 -- # [[ 0x0010 == \0\x\0\a\5\4 ]] 00:05:41.773 15:10:09 -- common/autotest_common.sh@1565 -- # for bdf in "${_bdfs[@]}" 00:05:41.773 15:10:09 -- common/autotest_common.sh@1566 -- # cat /sys/bus/pci/devices/0000:00:11.0/device 00:05:41.773 15:10:09 -- common/autotest_common.sh@1566 -- # device=0x0010 00:05:41.773 15:10:09 -- common/autotest_common.sh@1567 -- # [[ 0x0010 == \0\x\0\a\5\4 ]] 00:05:41.773 15:10:09 -- common/autotest_common.sh@1572 -- # (( 0 > 0 )) 00:05:41.773 15:10:09 -- common/autotest_common.sh@1572 -- # return 0 00:05:41.773 15:10:09 -- common/autotest_common.sh@1579 -- # [[ -z '' ]] 00:05:41.773 15:10:09 -- common/autotest_common.sh@1580 -- # return 0 00:05:41.773 15:10:09 -- spdk/autotest.sh@137 -- # '[' 0 -eq 1 ']' 00:05:41.773 15:10:09 -- spdk/autotest.sh@141 -- # '[' 1 -eq 1 ']' 00:05:41.773 15:10:09 -- spdk/autotest.sh@142 -- # [[ 0 -eq 1 ]] 00:05:41.773 15:10:09 -- spdk/autotest.sh@142 -- # [[ 0 -eq 1 ]] 00:05:41.773 15:10:09 -- spdk/autotest.sh@149 -- # timing_enter lib 00:05:41.773 15:10:09 -- common/autotest_common.sh@726 -- # xtrace_disable 00:05:41.773 15:10:09 -- common/autotest_common.sh@10 -- # set +x 00:05:41.773 15:10:09 -- spdk/autotest.sh@151 -- # [[ 0 -eq 1 ]] 00:05:41.773 15:10:09 -- spdk/autotest.sh@155 -- # run_test env /home/vagrant/spdk_repo/spdk/test/env/env.sh 00:05:41.773 15:10:09 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:41.773 15:10:09 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:41.773 15:10:09 -- common/autotest_common.sh@10 -- # set +x 00:05:41.773 ************************************ 00:05:41.773 START TEST env 00:05:41.773 ************************************ 00:05:41.773 15:10:09 env -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/env.sh 00:05:42.033 * Looking for test storage... 00:05:42.033 * Found test storage at /home/vagrant/spdk_repo/spdk/test/env 00:05:42.033 15:10:09 env -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:05:42.033 15:10:09 env -- common/autotest_common.sh@1693 -- # lcov --version 00:05:42.033 15:10:09 env -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:05:42.033 15:10:09 env -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:05:42.033 15:10:09 env -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:42.033 15:10:09 env -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:42.033 15:10:09 env -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:42.033 15:10:09 env -- scripts/common.sh@336 -- # IFS=.-: 00:05:42.033 15:10:09 env -- scripts/common.sh@336 -- # read -ra ver1 00:05:42.033 15:10:09 env -- scripts/common.sh@337 -- # IFS=.-: 00:05:42.033 15:10:09 env -- scripts/common.sh@337 -- # read -ra ver2 00:05:42.033 15:10:09 env -- scripts/common.sh@338 -- # local 'op=<' 00:05:42.033 15:10:09 env -- scripts/common.sh@340 -- # ver1_l=2 00:05:42.033 15:10:09 env -- scripts/common.sh@341 -- # ver2_l=1 00:05:42.033 15:10:09 env -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:42.033 15:10:09 env -- scripts/common.sh@344 -- # case "$op" in 00:05:42.033 15:10:09 env -- scripts/common.sh@345 -- # : 1 00:05:42.033 15:10:09 env -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:42.033 15:10:09 env -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:42.033 15:10:09 env -- scripts/common.sh@365 -- # decimal 1 00:05:42.033 15:10:09 env -- scripts/common.sh@353 -- # local d=1 00:05:42.033 15:10:09 env -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:42.033 15:10:09 env -- scripts/common.sh@355 -- # echo 1 00:05:42.033 15:10:09 env -- scripts/common.sh@365 -- # ver1[v]=1 00:05:42.033 15:10:09 env -- scripts/common.sh@366 -- # decimal 2 00:05:42.033 15:10:09 env -- scripts/common.sh@353 -- # local d=2 00:05:42.033 15:10:09 env -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:42.033 15:10:09 env -- scripts/common.sh@355 -- # echo 2 00:05:42.033 15:10:09 env -- scripts/common.sh@366 -- # ver2[v]=2 00:05:42.033 15:10:09 env -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:42.033 15:10:09 env -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:42.033 15:10:09 env -- scripts/common.sh@368 -- # return 0 00:05:42.033 15:10:09 env -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:42.033 15:10:09 env -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:05:42.033 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:42.033 --rc genhtml_branch_coverage=1 00:05:42.033 --rc genhtml_function_coverage=1 00:05:42.033 --rc genhtml_legend=1 00:05:42.033 --rc geninfo_all_blocks=1 00:05:42.033 --rc geninfo_unexecuted_blocks=1 00:05:42.033 00:05:42.033 ' 00:05:42.033 15:10:09 env -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:05:42.033 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:42.033 --rc genhtml_branch_coverage=1 00:05:42.033 --rc genhtml_function_coverage=1 00:05:42.033 --rc genhtml_legend=1 00:05:42.033 --rc geninfo_all_blocks=1 00:05:42.033 --rc geninfo_unexecuted_blocks=1 00:05:42.033 00:05:42.033 ' 00:05:42.033 15:10:09 env -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:05:42.033 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:42.033 --rc genhtml_branch_coverage=1 00:05:42.033 --rc genhtml_function_coverage=1 00:05:42.033 --rc genhtml_legend=1 00:05:42.033 --rc geninfo_all_blocks=1 00:05:42.033 --rc geninfo_unexecuted_blocks=1 00:05:42.033 00:05:42.033 ' 00:05:42.033 15:10:09 env -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:05:42.033 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:42.033 --rc genhtml_branch_coverage=1 00:05:42.033 --rc genhtml_function_coverage=1 00:05:42.033 --rc genhtml_legend=1 00:05:42.033 --rc geninfo_all_blocks=1 00:05:42.033 --rc geninfo_unexecuted_blocks=1 00:05:42.033 00:05:42.033 ' 00:05:42.033 15:10:09 env -- env/env.sh@10 -- # run_test env_memory /home/vagrant/spdk_repo/spdk/test/env/memory/memory_ut 00:05:42.033 15:10:09 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:42.033 15:10:09 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:42.033 15:10:09 env -- common/autotest_common.sh@10 -- # set +x 00:05:42.033 ************************************ 00:05:42.033 START TEST env_memory 00:05:42.033 ************************************ 00:05:42.033 15:10:10 env.env_memory -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/memory/memory_ut 00:05:42.033 00:05:42.033 00:05:42.033 CUnit - A unit testing framework for C - Version 2.1-3 00:05:42.033 http://cunit.sourceforge.net/ 00:05:42.033 00:05:42.033 00:05:42.033 Suite: memory 00:05:42.033 Test: alloc and free memory map ...[2024-11-27 15:10:10.075725] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 283:spdk_mem_map_alloc: *ERROR*: Initial mem_map notify failed 00:05:42.033 passed 00:05:42.033 Test: mem map translation ...[2024-11-27 15:10:10.119872] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 595:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=2097152 len=1234 00:05:42.033 [2024-11-27 15:10:10.119944] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 595:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=1234 len=2097152 00:05:42.033 [2024-11-27 15:10:10.120022] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 589:spdk_mem_map_set_translation: *ERROR*: invalid usermode virtual address 281474976710656 00:05:42.033 [2024-11-27 15:10:10.120045] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 605:spdk_mem_map_set_translation: *ERROR*: could not get 0xffffffe00000 map 00:05:42.293 passed 00:05:42.293 Test: mem map registration ...[2024-11-27 15:10:10.193484] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 347:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=200000 len=1234 00:05:42.293 [2024-11-27 15:10:10.193530] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 347:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=4d2 len=2097152 00:05:42.293 passed 00:05:42.293 Test: mem map adjacent registrations ...passed 00:05:42.293 00:05:42.293 Run Summary: Type Total Ran Passed Failed Inactive 00:05:42.293 suites 1 1 n/a 0 0 00:05:42.293 tests 4 4 4 0 0 00:05:42.293 asserts 152 152 152 0 n/a 00:05:42.293 00:05:42.293 Elapsed time = 0.247 seconds 00:05:42.293 00:05:42.293 real 0m0.300s 00:05:42.293 user 0m0.257s 00:05:42.293 sys 0m0.032s 00:05:42.293 15:10:10 env.env_memory -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:42.293 15:10:10 env.env_memory -- common/autotest_common.sh@10 -- # set +x 00:05:42.293 ************************************ 00:05:42.293 END TEST env_memory 00:05:42.293 ************************************ 00:05:42.293 15:10:10 env -- env/env.sh@11 -- # run_test env_vtophys /home/vagrant/spdk_repo/spdk/test/env/vtophys/vtophys 00:05:42.293 15:10:10 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:42.293 15:10:10 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:42.293 15:10:10 env -- common/autotest_common.sh@10 -- # set +x 00:05:42.293 ************************************ 00:05:42.293 START TEST env_vtophys 00:05:42.293 ************************************ 00:05:42.293 15:10:10 env.env_vtophys -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/vtophys/vtophys 00:05:42.553 EAL: lib.eal log level changed from notice to debug 00:05:42.553 EAL: Detected lcore 0 as core 0 on socket 0 00:05:42.553 EAL: Detected lcore 1 as core 0 on socket 0 00:05:42.554 EAL: Detected lcore 2 as core 0 on socket 0 00:05:42.554 EAL: Detected lcore 3 as core 0 on socket 0 00:05:42.554 EAL: Detected lcore 4 as core 0 on socket 0 00:05:42.554 EAL: Detected lcore 5 as core 0 on socket 0 00:05:42.554 EAL: Detected lcore 6 as core 0 on socket 0 00:05:42.554 EAL: Detected lcore 7 as core 0 on socket 0 00:05:42.554 EAL: Detected lcore 8 as core 0 on socket 0 00:05:42.554 EAL: Detected lcore 9 as core 0 on socket 0 00:05:42.554 EAL: Maximum logical cores by configuration: 128 00:05:42.554 EAL: Detected CPU lcores: 10 00:05:42.554 EAL: Detected NUMA nodes: 1 00:05:42.554 EAL: Checking presence of .so 'librte_eal.so.24.0' 00:05:42.554 EAL: Detected shared linkage of DPDK 00:05:42.554 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0/librte_bus_pci.so.24.0 00:05:42.554 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0/librte_bus_vdev.so.24.0 00:05:42.554 EAL: Registered [vdev] bus. 00:05:42.554 EAL: bus.vdev log level changed from disabled to notice 00:05:42.554 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0/librte_mempool_ring.so.24.0 00:05:42.554 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0/librte_net_i40e.so.24.0 00:05:42.554 EAL: pmd.net.i40e.init log level changed from disabled to notice 00:05:42.554 EAL: pmd.net.i40e.driver log level changed from disabled to notice 00:05:42.554 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0/librte_bus_pci.so 00:05:42.554 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0/librte_bus_vdev.so 00:05:42.554 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0/librte_mempool_ring.so 00:05:42.554 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0/librte_net_i40e.so 00:05:42.554 EAL: No shared files mode enabled, IPC will be disabled 00:05:42.554 EAL: No shared files mode enabled, IPC is disabled 00:05:42.554 EAL: Selected IOVA mode 'PA' 00:05:42.554 EAL: Probing VFIO support... 00:05:42.554 EAL: Module /sys/module/vfio not found! error 2 (No such file or directory) 00:05:42.554 EAL: VFIO modules not loaded, skipping VFIO support... 00:05:42.554 EAL: Ask a virtual area of 0x2e000 bytes 00:05:42.554 EAL: Virtual area found at 0x200000000000 (size = 0x2e000) 00:05:42.554 EAL: Setting up physically contiguous memory... 00:05:42.554 EAL: Setting maximum number of open files to 524288 00:05:42.554 EAL: Detected memory type: socket_id:0 hugepage_sz:2097152 00:05:42.554 EAL: Creating 4 segment lists: n_segs:8192 socket_id:0 hugepage_sz:2097152 00:05:42.554 EAL: Ask a virtual area of 0x61000 bytes 00:05:42.554 EAL: Virtual area found at 0x20000002e000 (size = 0x61000) 00:05:42.554 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:05:42.554 EAL: Ask a virtual area of 0x400000000 bytes 00:05:42.554 EAL: Virtual area found at 0x200000200000 (size = 0x400000000) 00:05:42.554 EAL: VA reserved for memseg list at 0x200000200000, size 400000000 00:05:42.554 EAL: Ask a virtual area of 0x61000 bytes 00:05:42.554 EAL: Virtual area found at 0x200400200000 (size = 0x61000) 00:05:42.554 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:05:42.554 EAL: Ask a virtual area of 0x400000000 bytes 00:05:42.554 EAL: Virtual area found at 0x200400400000 (size = 0x400000000) 00:05:42.554 EAL: VA reserved for memseg list at 0x200400400000, size 400000000 00:05:42.554 EAL: Ask a virtual area of 0x61000 bytes 00:05:42.554 EAL: Virtual area found at 0x200800400000 (size = 0x61000) 00:05:42.554 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:05:42.554 EAL: Ask a virtual area of 0x400000000 bytes 00:05:42.554 EAL: Virtual area found at 0x200800600000 (size = 0x400000000) 00:05:42.554 EAL: VA reserved for memseg list at 0x200800600000, size 400000000 00:05:42.554 EAL: Ask a virtual area of 0x61000 bytes 00:05:42.554 EAL: Virtual area found at 0x200c00600000 (size = 0x61000) 00:05:42.554 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:05:42.554 EAL: Ask a virtual area of 0x400000000 bytes 00:05:42.554 EAL: Virtual area found at 0x200c00800000 (size = 0x400000000) 00:05:42.554 EAL: VA reserved for memseg list at 0x200c00800000, size 400000000 00:05:42.554 EAL: Hugepages will be freed exactly as allocated. 00:05:42.554 EAL: No shared files mode enabled, IPC is disabled 00:05:42.554 EAL: No shared files mode enabled, IPC is disabled 00:05:42.554 EAL: TSC frequency is ~2290000 KHz 00:05:42.554 EAL: Main lcore 0 is ready (tid=7f1783a89a40;cpuset=[0]) 00:05:42.554 EAL: Trying to obtain current memory policy. 00:05:42.554 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:42.554 EAL: Restoring previous memory policy: 0 00:05:42.554 EAL: request: mp_malloc_sync 00:05:42.554 EAL: No shared files mode enabled, IPC is disabled 00:05:42.554 EAL: Heap on socket 0 was expanded by 2MB 00:05:42.554 EAL: Module /sys/module/vfio not found! error 2 (No such file or directory) 00:05:42.554 EAL: No shared files mode enabled, IPC is disabled 00:05:42.554 EAL: No PCI address specified using 'addr=' in: bus=pci 00:05:42.554 EAL: Mem event callback 'spdk:(nil)' registered 00:05:42.554 EAL: Module /sys/module/vfio_pci not found! error 2 (No such file or directory) 00:05:42.554 00:05:42.554 00:05:42.554 CUnit - A unit testing framework for C - Version 2.1-3 00:05:42.554 http://cunit.sourceforge.net/ 00:05:42.554 00:05:42.554 00:05:42.554 Suite: components_suite 00:05:43.161 Test: vtophys_malloc_test ...passed 00:05:43.161 Test: vtophys_spdk_malloc_test ...EAL: Trying to obtain current memory policy. 00:05:43.161 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:43.161 EAL: Restoring previous memory policy: 4 00:05:43.161 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.161 EAL: request: mp_malloc_sync 00:05:43.161 EAL: No shared files mode enabled, IPC is disabled 00:05:43.161 EAL: Heap on socket 0 was expanded by 4MB 00:05:43.161 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.161 EAL: request: mp_malloc_sync 00:05:43.161 EAL: No shared files mode enabled, IPC is disabled 00:05:43.161 EAL: Heap on socket 0 was shrunk by 4MB 00:05:43.161 EAL: Trying to obtain current memory policy. 00:05:43.161 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:43.161 EAL: Restoring previous memory policy: 4 00:05:43.161 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.161 EAL: request: mp_malloc_sync 00:05:43.161 EAL: No shared files mode enabled, IPC is disabled 00:05:43.161 EAL: Heap on socket 0 was expanded by 6MB 00:05:43.161 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.161 EAL: request: mp_malloc_sync 00:05:43.161 EAL: No shared files mode enabled, IPC is disabled 00:05:43.161 EAL: Heap on socket 0 was shrunk by 6MB 00:05:43.161 EAL: Trying to obtain current memory policy. 00:05:43.161 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:43.161 EAL: Restoring previous memory policy: 4 00:05:43.161 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.161 EAL: request: mp_malloc_sync 00:05:43.161 EAL: No shared files mode enabled, IPC is disabled 00:05:43.161 EAL: Heap on socket 0 was expanded by 10MB 00:05:43.161 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.161 EAL: request: mp_malloc_sync 00:05:43.161 EAL: No shared files mode enabled, IPC is disabled 00:05:43.161 EAL: Heap on socket 0 was shrunk by 10MB 00:05:43.161 EAL: Trying to obtain current memory policy. 00:05:43.161 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:43.161 EAL: Restoring previous memory policy: 4 00:05:43.161 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.161 EAL: request: mp_malloc_sync 00:05:43.161 EAL: No shared files mode enabled, IPC is disabled 00:05:43.161 EAL: Heap on socket 0 was expanded by 18MB 00:05:43.161 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.161 EAL: request: mp_malloc_sync 00:05:43.161 EAL: No shared files mode enabled, IPC is disabled 00:05:43.161 EAL: Heap on socket 0 was shrunk by 18MB 00:05:43.161 EAL: Trying to obtain current memory policy. 00:05:43.161 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:43.161 EAL: Restoring previous memory policy: 4 00:05:43.161 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.161 EAL: request: mp_malloc_sync 00:05:43.161 EAL: No shared files mode enabled, IPC is disabled 00:05:43.161 EAL: Heap on socket 0 was expanded by 34MB 00:05:43.161 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.161 EAL: request: mp_malloc_sync 00:05:43.161 EAL: No shared files mode enabled, IPC is disabled 00:05:43.161 EAL: Heap on socket 0 was shrunk by 34MB 00:05:43.161 EAL: Trying to obtain current memory policy. 00:05:43.161 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:43.161 EAL: Restoring previous memory policy: 4 00:05:43.161 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.161 EAL: request: mp_malloc_sync 00:05:43.161 EAL: No shared files mode enabled, IPC is disabled 00:05:43.161 EAL: Heap on socket 0 was expanded by 66MB 00:05:43.161 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.161 EAL: request: mp_malloc_sync 00:05:43.161 EAL: No shared files mode enabled, IPC is disabled 00:05:43.161 EAL: Heap on socket 0 was shrunk by 66MB 00:05:43.161 EAL: Trying to obtain current memory policy. 00:05:43.161 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:43.161 EAL: Restoring previous memory policy: 4 00:05:43.161 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.161 EAL: request: mp_malloc_sync 00:05:43.161 EAL: No shared files mode enabled, IPC is disabled 00:05:43.161 EAL: Heap on socket 0 was expanded by 130MB 00:05:43.161 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.161 EAL: request: mp_malloc_sync 00:05:43.161 EAL: No shared files mode enabled, IPC is disabled 00:05:43.161 EAL: Heap on socket 0 was shrunk by 130MB 00:05:43.161 EAL: Trying to obtain current memory policy. 00:05:43.161 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:43.161 EAL: Restoring previous memory policy: 4 00:05:43.161 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.161 EAL: request: mp_malloc_sync 00:05:43.161 EAL: No shared files mode enabled, IPC is disabled 00:05:43.161 EAL: Heap on socket 0 was expanded by 258MB 00:05:43.161 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.161 EAL: request: mp_malloc_sync 00:05:43.161 EAL: No shared files mode enabled, IPC is disabled 00:05:43.161 EAL: Heap on socket 0 was shrunk by 258MB 00:05:43.161 EAL: Trying to obtain current memory policy. 00:05:43.161 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:43.424 EAL: Restoring previous memory policy: 4 00:05:43.424 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.424 EAL: request: mp_malloc_sync 00:05:43.424 EAL: No shared files mode enabled, IPC is disabled 00:05:43.424 EAL: Heap on socket 0 was expanded by 514MB 00:05:43.424 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.424 EAL: request: mp_malloc_sync 00:05:43.424 EAL: No shared files mode enabled, IPC is disabled 00:05:43.424 EAL: Heap on socket 0 was shrunk by 514MB 00:05:43.424 EAL: Trying to obtain current memory policy. 00:05:43.424 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:43.684 EAL: Restoring previous memory policy: 4 00:05:43.684 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.684 EAL: request: mp_malloc_sync 00:05:43.684 EAL: No shared files mode enabled, IPC is disabled 00:05:43.684 EAL: Heap on socket 0 was expanded by 1026MB 00:05:43.945 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.945 passed 00:05:43.945 00:05:43.945 Run Summary: Type Total Ran Passed Failed Inactive 00:05:43.945 suites 1 1 n/a 0 0 00:05:43.945 tests 2 2 2 0 0 00:05:43.945 asserts 5512 5512 5512 0 n/a 00:05:43.945 00:05:43.945 Elapsed time = 1.342 seconds 00:05:43.945 EAL: request: mp_malloc_sync 00:05:43.945 EAL: No shared files mode enabled, IPC is disabled 00:05:43.945 EAL: Heap on socket 0 was shrunk by 1026MB 00:05:43.945 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.945 EAL: request: mp_malloc_sync 00:05:43.945 EAL: No shared files mode enabled, IPC is disabled 00:05:43.945 EAL: Heap on socket 0 was shrunk by 2MB 00:05:43.945 EAL: No shared files mode enabled, IPC is disabled 00:05:43.945 EAL: No shared files mode enabled, IPC is disabled 00:05:43.945 EAL: No shared files mode enabled, IPC is disabled 00:05:43.945 ************************************ 00:05:43.945 END TEST env_vtophys 00:05:43.945 ************************************ 00:05:43.945 00:05:43.945 real 0m1.620s 00:05:43.945 user 0m0.795s 00:05:43.945 sys 0m0.692s 00:05:43.945 15:10:11 env.env_vtophys -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:43.945 15:10:11 env.env_vtophys -- common/autotest_common.sh@10 -- # set +x 00:05:43.945 15:10:12 env -- env/env.sh@12 -- # run_test env_pci /home/vagrant/spdk_repo/spdk/test/env/pci/pci_ut 00:05:43.945 15:10:12 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:43.945 15:10:12 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:43.945 15:10:12 env -- common/autotest_common.sh@10 -- # set +x 00:05:44.205 ************************************ 00:05:44.205 START TEST env_pci 00:05:44.205 ************************************ 00:05:44.205 15:10:12 env.env_pci -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/pci/pci_ut 00:05:44.205 00:05:44.205 00:05:44.205 CUnit - A unit testing framework for C - Version 2.1-3 00:05:44.205 http://cunit.sourceforge.net/ 00:05:44.205 00:05:44.205 00:05:44.205 Suite: pci 00:05:44.205 Test: pci_hook ...[2024-11-27 15:10:12.088134] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/pci.c:1117:spdk_pci_device_claim: *ERROR*: Cannot create lock on device /var/tmp/spdk_pci_lock_10000:00:01.0, probably process 69174 has claimed it 00:05:44.205 EAL: Cannot find device (10000:00:01.0) 00:05:44.205 EAL: Failed to attach device on primary process 00:05:44.205 passed 00:05:44.205 00:05:44.205 Run Summary: Type Total Ran Passed Failed Inactive 00:05:44.205 suites 1 1 n/a 0 0 00:05:44.205 tests 1 1 1 0 0 00:05:44.205 asserts 25 25 25 0 n/a 00:05:44.205 00:05:44.205 Elapsed time = 0.006 seconds 00:05:44.205 00:05:44.205 real 0m0.094s 00:05:44.205 user 0m0.041s 00:05:44.205 sys 0m0.051s 00:05:44.205 ************************************ 00:05:44.205 END TEST env_pci 00:05:44.205 ************************************ 00:05:44.205 15:10:12 env.env_pci -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:44.205 15:10:12 env.env_pci -- common/autotest_common.sh@10 -- # set +x 00:05:44.205 15:10:12 env -- env/env.sh@14 -- # argv='-c 0x1 ' 00:05:44.205 15:10:12 env -- env/env.sh@15 -- # uname 00:05:44.205 15:10:12 env -- env/env.sh@15 -- # '[' Linux = Linux ']' 00:05:44.205 15:10:12 env -- env/env.sh@22 -- # argv+=--base-virtaddr=0x200000000000 00:05:44.205 15:10:12 env -- env/env.sh@24 -- # run_test env_dpdk_post_init /home/vagrant/spdk_repo/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:05:44.205 15:10:12 env -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:05:44.205 15:10:12 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:44.205 15:10:12 env -- common/autotest_common.sh@10 -- # set +x 00:05:44.205 ************************************ 00:05:44.205 START TEST env_dpdk_post_init 00:05:44.205 ************************************ 00:05:44.205 15:10:12 env.env_dpdk_post_init -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:05:44.205 EAL: Detected CPU lcores: 10 00:05:44.205 EAL: Detected NUMA nodes: 1 00:05:44.205 EAL: Detected shared linkage of DPDK 00:05:44.205 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:05:44.205 EAL: Selected IOVA mode 'PA' 00:05:44.465 TELEMETRY: No legacy callbacks, legacy socket not created 00:05:44.465 EAL: Probe PCI driver: spdk_nvme (1b36:0010) device: 0000:00:10.0 (socket -1) 00:05:44.465 EAL: Probe PCI driver: spdk_nvme (1b36:0010) device: 0000:00:11.0 (socket -1) 00:05:44.465 Starting DPDK initialization... 00:05:44.465 Starting SPDK post initialization... 00:05:44.465 SPDK NVMe probe 00:05:44.465 Attaching to 0000:00:10.0 00:05:44.465 Attaching to 0000:00:11.0 00:05:44.465 Attached to 0000:00:10.0 00:05:44.465 Attached to 0000:00:11.0 00:05:44.465 Cleaning up... 00:05:44.465 ************************************ 00:05:44.465 END TEST env_dpdk_post_init 00:05:44.465 ************************************ 00:05:44.465 00:05:44.465 real 0m0.263s 00:05:44.465 user 0m0.082s 00:05:44.465 sys 0m0.082s 00:05:44.465 15:10:12 env.env_dpdk_post_init -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:44.465 15:10:12 env.env_dpdk_post_init -- common/autotest_common.sh@10 -- # set +x 00:05:44.465 15:10:12 env -- env/env.sh@26 -- # uname 00:05:44.465 15:10:12 env -- env/env.sh@26 -- # '[' Linux = Linux ']' 00:05:44.465 15:10:12 env -- env/env.sh@29 -- # run_test env_mem_callbacks /home/vagrant/spdk_repo/spdk/test/env/mem_callbacks/mem_callbacks 00:05:44.465 15:10:12 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:44.465 15:10:12 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:44.465 15:10:12 env -- common/autotest_common.sh@10 -- # set +x 00:05:44.465 ************************************ 00:05:44.465 START TEST env_mem_callbacks 00:05:44.465 ************************************ 00:05:44.465 15:10:12 env.env_mem_callbacks -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/mem_callbacks/mem_callbacks 00:05:44.725 EAL: Detected CPU lcores: 10 00:05:44.725 EAL: Detected NUMA nodes: 1 00:05:44.725 EAL: Detected shared linkage of DPDK 00:05:44.725 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:05:44.725 EAL: Selected IOVA mode 'PA' 00:05:44.725 TELEMETRY: No legacy callbacks, legacy socket not created 00:05:44.725 00:05:44.725 00:05:44.725 CUnit - A unit testing framework for C - Version 2.1-3 00:05:44.725 http://cunit.sourceforge.net/ 00:05:44.725 00:05:44.725 00:05:44.725 Suite: memory 00:05:44.725 Test: test ... 00:05:44.725 register 0x200000200000 2097152 00:05:44.725 malloc 3145728 00:05:44.725 register 0x200000400000 4194304 00:05:44.725 buf 0x200000500000 len 3145728 PASSED 00:05:44.725 malloc 64 00:05:44.725 buf 0x2000004fff40 len 64 PASSED 00:05:44.725 malloc 4194304 00:05:44.725 register 0x200000800000 6291456 00:05:44.725 buf 0x200000a00000 len 4194304 PASSED 00:05:44.725 free 0x200000500000 3145728 00:05:44.725 free 0x2000004fff40 64 00:05:44.725 unregister 0x200000400000 4194304 PASSED 00:05:44.725 free 0x200000a00000 4194304 00:05:44.725 unregister 0x200000800000 6291456 PASSED 00:05:44.725 malloc 8388608 00:05:44.725 register 0x200000400000 10485760 00:05:44.725 buf 0x200000600000 len 8388608 PASSED 00:05:44.725 free 0x200000600000 8388608 00:05:44.725 unregister 0x200000400000 10485760 PASSED 00:05:44.725 passed 00:05:44.725 00:05:44.725 Run Summary: Type Total Ran Passed Failed Inactive 00:05:44.725 suites 1 1 n/a 0 0 00:05:44.725 tests 1 1 1 0 0 00:05:44.725 asserts 15 15 15 0 n/a 00:05:44.725 00:05:44.725 Elapsed time = 0.012 seconds 00:05:44.725 00:05:44.725 real 0m0.211s 00:05:44.725 user 0m0.031s 00:05:44.725 sys 0m0.076s 00:05:44.725 15:10:12 env.env_mem_callbacks -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:44.725 15:10:12 env.env_mem_callbacks -- common/autotest_common.sh@10 -- # set +x 00:05:44.725 ************************************ 00:05:44.725 END TEST env_mem_callbacks 00:05:44.725 ************************************ 00:05:44.725 ************************************ 00:05:44.725 END TEST env 00:05:44.725 ************************************ 00:05:44.725 00:05:44.725 real 0m3.055s 00:05:44.725 user 0m1.416s 00:05:44.725 sys 0m1.308s 00:05:44.725 15:10:12 env -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:44.725 15:10:12 env -- common/autotest_common.sh@10 -- # set +x 00:05:44.985 15:10:12 -- spdk/autotest.sh@156 -- # run_test rpc /home/vagrant/spdk_repo/spdk/test/rpc/rpc.sh 00:05:44.985 15:10:12 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:44.985 15:10:12 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:44.985 15:10:12 -- common/autotest_common.sh@10 -- # set +x 00:05:44.985 ************************************ 00:05:44.985 START TEST rpc 00:05:44.985 ************************************ 00:05:44.985 15:10:12 rpc -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/rpc/rpc.sh 00:05:44.985 * Looking for test storage... 00:05:44.985 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc 00:05:44.985 15:10:13 rpc -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:05:44.985 15:10:13 rpc -- common/autotest_common.sh@1693 -- # lcov --version 00:05:44.985 15:10:13 rpc -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:05:45.245 15:10:13 rpc -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:05:45.245 15:10:13 rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:45.245 15:10:13 rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:45.245 15:10:13 rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:45.245 15:10:13 rpc -- scripts/common.sh@336 -- # IFS=.-: 00:05:45.245 15:10:13 rpc -- scripts/common.sh@336 -- # read -ra ver1 00:05:45.245 15:10:13 rpc -- scripts/common.sh@337 -- # IFS=.-: 00:05:45.245 15:10:13 rpc -- scripts/common.sh@337 -- # read -ra ver2 00:05:45.245 15:10:13 rpc -- scripts/common.sh@338 -- # local 'op=<' 00:05:45.245 15:10:13 rpc -- scripts/common.sh@340 -- # ver1_l=2 00:05:45.245 15:10:13 rpc -- scripts/common.sh@341 -- # ver2_l=1 00:05:45.245 15:10:13 rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:45.245 15:10:13 rpc -- scripts/common.sh@344 -- # case "$op" in 00:05:45.245 15:10:13 rpc -- scripts/common.sh@345 -- # : 1 00:05:45.245 15:10:13 rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:45.245 15:10:13 rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:45.245 15:10:13 rpc -- scripts/common.sh@365 -- # decimal 1 00:05:45.245 15:10:13 rpc -- scripts/common.sh@353 -- # local d=1 00:05:45.245 15:10:13 rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:45.245 15:10:13 rpc -- scripts/common.sh@355 -- # echo 1 00:05:45.245 15:10:13 rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:05:45.245 15:10:13 rpc -- scripts/common.sh@366 -- # decimal 2 00:05:45.245 15:10:13 rpc -- scripts/common.sh@353 -- # local d=2 00:05:45.245 15:10:13 rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:45.245 15:10:13 rpc -- scripts/common.sh@355 -- # echo 2 00:05:45.245 15:10:13 rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:05:45.245 15:10:13 rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:45.245 15:10:13 rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:45.245 15:10:13 rpc -- scripts/common.sh@368 -- # return 0 00:05:45.245 15:10:13 rpc -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:45.245 15:10:13 rpc -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:05:45.245 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:45.245 --rc genhtml_branch_coverage=1 00:05:45.245 --rc genhtml_function_coverage=1 00:05:45.245 --rc genhtml_legend=1 00:05:45.245 --rc geninfo_all_blocks=1 00:05:45.245 --rc geninfo_unexecuted_blocks=1 00:05:45.245 00:05:45.245 ' 00:05:45.245 15:10:13 rpc -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:05:45.245 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:45.245 --rc genhtml_branch_coverage=1 00:05:45.245 --rc genhtml_function_coverage=1 00:05:45.245 --rc genhtml_legend=1 00:05:45.245 --rc geninfo_all_blocks=1 00:05:45.245 --rc geninfo_unexecuted_blocks=1 00:05:45.245 00:05:45.245 ' 00:05:45.245 15:10:13 rpc -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:05:45.245 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:45.245 --rc genhtml_branch_coverage=1 00:05:45.245 --rc genhtml_function_coverage=1 00:05:45.245 --rc genhtml_legend=1 00:05:45.245 --rc geninfo_all_blocks=1 00:05:45.245 --rc geninfo_unexecuted_blocks=1 00:05:45.245 00:05:45.245 ' 00:05:45.245 15:10:13 rpc -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:05:45.245 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:45.245 --rc genhtml_branch_coverage=1 00:05:45.245 --rc genhtml_function_coverage=1 00:05:45.245 --rc genhtml_legend=1 00:05:45.245 --rc geninfo_all_blocks=1 00:05:45.245 --rc geninfo_unexecuted_blocks=1 00:05:45.245 00:05:45.245 ' 00:05:45.245 15:10:13 rpc -- rpc/rpc.sh@65 -- # spdk_pid=69301 00:05:45.245 15:10:13 rpc -- rpc/rpc.sh@64 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -e bdev 00:05:45.245 15:10:13 rpc -- rpc/rpc.sh@66 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:05:45.245 15:10:13 rpc -- rpc/rpc.sh@67 -- # waitforlisten 69301 00:05:45.245 15:10:13 rpc -- common/autotest_common.sh@835 -- # '[' -z 69301 ']' 00:05:45.245 15:10:13 rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:45.245 15:10:13 rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:45.245 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:45.245 15:10:13 rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:45.245 15:10:13 rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:45.245 15:10:13 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:45.245 [2024-11-27 15:10:13.210531] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:05:45.245 [2024-11-27 15:10:13.210657] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69301 ] 00:05:45.506 [2024-11-27 15:10:13.383638] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:45.506 [2024-11-27 15:10:13.413312] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask bdev specified. 00:05:45.506 [2024-11-27 15:10:13.413480] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s spdk_tgt -p 69301' to capture a snapshot of events at runtime. 00:05:45.506 [2024-11-27 15:10:13.413497] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:05:45.506 [2024-11-27 15:10:13.413506] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:05:45.506 [2024-11-27 15:10:13.413516] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/spdk_tgt_trace.pid69301 for offline analysis/debug. 00:05:45.506 [2024-11-27 15:10:13.413892] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:46.077 15:10:14 rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:46.077 15:10:14 rpc -- common/autotest_common.sh@868 -- # return 0 00:05:46.077 15:10:14 rpc -- rpc/rpc.sh@69 -- # export PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/test/rpc 00:05:46.077 15:10:14 rpc -- rpc/rpc.sh@69 -- # PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/test/rpc 00:05:46.077 15:10:14 rpc -- rpc/rpc.sh@72 -- # rpc=rpc_cmd 00:05:46.077 15:10:14 rpc -- rpc/rpc.sh@73 -- # run_test rpc_integrity rpc_integrity 00:05:46.077 15:10:14 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:46.077 15:10:14 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:46.077 15:10:14 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:46.077 ************************************ 00:05:46.077 START TEST rpc_integrity 00:05:46.077 ************************************ 00:05:46.077 15:10:14 rpc.rpc_integrity -- common/autotest_common.sh@1129 -- # rpc_integrity 00:05:46.077 15:10:14 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:05:46.077 15:10:14 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:46.077 15:10:14 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:46.077 15:10:14 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:46.077 15:10:14 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:05:46.077 15:10:14 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # jq length 00:05:46.077 15:10:14 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:05:46.077 15:10:14 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:05:46.077 15:10:14 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:46.077 15:10:14 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:46.077 15:10:14 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:46.077 15:10:14 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc0 00:05:46.077 15:10:14 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:05:46.077 15:10:14 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:46.077 15:10:14 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:46.077 15:10:14 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:46.077 15:10:14 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:05:46.077 { 00:05:46.077 "name": "Malloc0", 00:05:46.077 "aliases": [ 00:05:46.077 "5c33b63e-93c9-41d3-ac57-89615d899c2a" 00:05:46.077 ], 00:05:46.077 "product_name": "Malloc disk", 00:05:46.077 "block_size": 512, 00:05:46.077 "num_blocks": 16384, 00:05:46.077 "uuid": "5c33b63e-93c9-41d3-ac57-89615d899c2a", 00:05:46.077 "assigned_rate_limits": { 00:05:46.077 "rw_ios_per_sec": 0, 00:05:46.077 "rw_mbytes_per_sec": 0, 00:05:46.077 "r_mbytes_per_sec": 0, 00:05:46.077 "w_mbytes_per_sec": 0 00:05:46.077 }, 00:05:46.077 "claimed": false, 00:05:46.077 "zoned": false, 00:05:46.077 "supported_io_types": { 00:05:46.077 "read": true, 00:05:46.077 "write": true, 00:05:46.077 "unmap": true, 00:05:46.077 "flush": true, 00:05:46.077 "reset": true, 00:05:46.077 "nvme_admin": false, 00:05:46.077 "nvme_io": false, 00:05:46.077 "nvme_io_md": false, 00:05:46.077 "write_zeroes": true, 00:05:46.077 "zcopy": true, 00:05:46.077 "get_zone_info": false, 00:05:46.077 "zone_management": false, 00:05:46.077 "zone_append": false, 00:05:46.077 "compare": false, 00:05:46.077 "compare_and_write": false, 00:05:46.077 "abort": true, 00:05:46.077 "seek_hole": false, 00:05:46.077 "seek_data": false, 00:05:46.077 "copy": true, 00:05:46.077 "nvme_iov_md": false 00:05:46.077 }, 00:05:46.077 "memory_domains": [ 00:05:46.077 { 00:05:46.077 "dma_device_id": "system", 00:05:46.077 "dma_device_type": 1 00:05:46.077 }, 00:05:46.077 { 00:05:46.077 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:46.077 "dma_device_type": 2 00:05:46.077 } 00:05:46.077 ], 00:05:46.077 "driver_specific": {} 00:05:46.077 } 00:05:46.077 ]' 00:05:46.077 15:10:14 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # jq length 00:05:46.338 15:10:14 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:05:46.338 15:10:14 rpc.rpc_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc0 -p Passthru0 00:05:46.338 15:10:14 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:46.338 15:10:14 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:46.338 [2024-11-27 15:10:14.236045] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on Malloc0 00:05:46.338 [2024-11-27 15:10:14.236142] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:05:46.338 [2024-11-27 15:10:14.236180] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:05:46.338 [2024-11-27 15:10:14.236192] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:05:46.338 [2024-11-27 15:10:14.238832] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:05:46.338 [2024-11-27 15:10:14.238886] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:05:46.338 Passthru0 00:05:46.338 15:10:14 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:46.338 15:10:14 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:05:46.338 15:10:14 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:46.338 15:10:14 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:46.338 15:10:14 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:46.338 15:10:14 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:05:46.338 { 00:05:46.338 "name": "Malloc0", 00:05:46.338 "aliases": [ 00:05:46.338 "5c33b63e-93c9-41d3-ac57-89615d899c2a" 00:05:46.338 ], 00:05:46.338 "product_name": "Malloc disk", 00:05:46.338 "block_size": 512, 00:05:46.338 "num_blocks": 16384, 00:05:46.338 "uuid": "5c33b63e-93c9-41d3-ac57-89615d899c2a", 00:05:46.338 "assigned_rate_limits": { 00:05:46.338 "rw_ios_per_sec": 0, 00:05:46.338 "rw_mbytes_per_sec": 0, 00:05:46.338 "r_mbytes_per_sec": 0, 00:05:46.338 "w_mbytes_per_sec": 0 00:05:46.338 }, 00:05:46.338 "claimed": true, 00:05:46.338 "claim_type": "exclusive_write", 00:05:46.338 "zoned": false, 00:05:46.338 "supported_io_types": { 00:05:46.338 "read": true, 00:05:46.338 "write": true, 00:05:46.338 "unmap": true, 00:05:46.338 "flush": true, 00:05:46.338 "reset": true, 00:05:46.338 "nvme_admin": false, 00:05:46.338 "nvme_io": false, 00:05:46.338 "nvme_io_md": false, 00:05:46.338 "write_zeroes": true, 00:05:46.338 "zcopy": true, 00:05:46.338 "get_zone_info": false, 00:05:46.338 "zone_management": false, 00:05:46.338 "zone_append": false, 00:05:46.338 "compare": false, 00:05:46.338 "compare_and_write": false, 00:05:46.338 "abort": true, 00:05:46.338 "seek_hole": false, 00:05:46.338 "seek_data": false, 00:05:46.338 "copy": true, 00:05:46.338 "nvme_iov_md": false 00:05:46.338 }, 00:05:46.338 "memory_domains": [ 00:05:46.338 { 00:05:46.338 "dma_device_id": "system", 00:05:46.338 "dma_device_type": 1 00:05:46.338 }, 00:05:46.338 { 00:05:46.338 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:46.338 "dma_device_type": 2 00:05:46.338 } 00:05:46.338 ], 00:05:46.338 "driver_specific": {} 00:05:46.338 }, 00:05:46.338 { 00:05:46.338 "name": "Passthru0", 00:05:46.338 "aliases": [ 00:05:46.338 "cd36abb8-7652-586f-9850-5fb51c8dbbe4" 00:05:46.338 ], 00:05:46.338 "product_name": "passthru", 00:05:46.338 "block_size": 512, 00:05:46.338 "num_blocks": 16384, 00:05:46.338 "uuid": "cd36abb8-7652-586f-9850-5fb51c8dbbe4", 00:05:46.338 "assigned_rate_limits": { 00:05:46.338 "rw_ios_per_sec": 0, 00:05:46.338 "rw_mbytes_per_sec": 0, 00:05:46.338 "r_mbytes_per_sec": 0, 00:05:46.338 "w_mbytes_per_sec": 0 00:05:46.338 }, 00:05:46.338 "claimed": false, 00:05:46.338 "zoned": false, 00:05:46.338 "supported_io_types": { 00:05:46.338 "read": true, 00:05:46.338 "write": true, 00:05:46.338 "unmap": true, 00:05:46.338 "flush": true, 00:05:46.338 "reset": true, 00:05:46.338 "nvme_admin": false, 00:05:46.338 "nvme_io": false, 00:05:46.338 "nvme_io_md": false, 00:05:46.338 "write_zeroes": true, 00:05:46.338 "zcopy": true, 00:05:46.338 "get_zone_info": false, 00:05:46.338 "zone_management": false, 00:05:46.338 "zone_append": false, 00:05:46.338 "compare": false, 00:05:46.338 "compare_and_write": false, 00:05:46.338 "abort": true, 00:05:46.338 "seek_hole": false, 00:05:46.338 "seek_data": false, 00:05:46.338 "copy": true, 00:05:46.338 "nvme_iov_md": false 00:05:46.338 }, 00:05:46.338 "memory_domains": [ 00:05:46.338 { 00:05:46.338 "dma_device_id": "system", 00:05:46.338 "dma_device_type": 1 00:05:46.338 }, 00:05:46.338 { 00:05:46.338 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:46.338 "dma_device_type": 2 00:05:46.338 } 00:05:46.338 ], 00:05:46.338 "driver_specific": { 00:05:46.338 "passthru": { 00:05:46.338 "name": "Passthru0", 00:05:46.338 "base_bdev_name": "Malloc0" 00:05:46.338 } 00:05:46.338 } 00:05:46.338 } 00:05:46.338 ]' 00:05:46.338 15:10:14 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # jq length 00:05:46.338 15:10:14 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:05:46.338 15:10:14 rpc.rpc_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:05:46.338 15:10:14 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:46.338 15:10:14 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:46.338 15:10:14 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:46.338 15:10:14 rpc.rpc_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc0 00:05:46.338 15:10:14 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:46.338 15:10:14 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:46.338 15:10:14 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:46.338 15:10:14 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:05:46.338 15:10:14 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:46.338 15:10:14 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:46.338 15:10:14 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:46.338 15:10:14 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:05:46.339 15:10:14 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # jq length 00:05:46.339 15:10:14 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:05:46.339 00:05:46.339 real 0m0.337s 00:05:46.339 user 0m0.190s 00:05:46.339 sys 0m0.069s 00:05:46.339 15:10:14 rpc.rpc_integrity -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:46.339 15:10:14 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:46.339 ************************************ 00:05:46.339 END TEST rpc_integrity 00:05:46.339 ************************************ 00:05:46.599 15:10:14 rpc -- rpc/rpc.sh@74 -- # run_test rpc_plugins rpc_plugins 00:05:46.599 15:10:14 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:46.599 15:10:14 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:46.599 15:10:14 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:46.599 ************************************ 00:05:46.599 START TEST rpc_plugins 00:05:46.599 ************************************ 00:05:46.599 15:10:14 rpc.rpc_plugins -- common/autotest_common.sh@1129 -- # rpc_plugins 00:05:46.599 15:10:14 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # rpc_cmd --plugin rpc_plugin create_malloc 00:05:46.599 15:10:14 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:46.599 15:10:14 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:05:46.599 15:10:14 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:46.599 15:10:14 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # malloc=Malloc1 00:05:46.599 15:10:14 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # rpc_cmd bdev_get_bdevs 00:05:46.599 15:10:14 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:46.599 15:10:14 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:05:46.599 15:10:14 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:46.599 15:10:14 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # bdevs='[ 00:05:46.599 { 00:05:46.599 "name": "Malloc1", 00:05:46.599 "aliases": [ 00:05:46.599 "9c2ea8e0-261a-4476-92ff-5f803c594153" 00:05:46.599 ], 00:05:46.599 "product_name": "Malloc disk", 00:05:46.599 "block_size": 4096, 00:05:46.599 "num_blocks": 256, 00:05:46.599 "uuid": "9c2ea8e0-261a-4476-92ff-5f803c594153", 00:05:46.599 "assigned_rate_limits": { 00:05:46.599 "rw_ios_per_sec": 0, 00:05:46.599 "rw_mbytes_per_sec": 0, 00:05:46.599 "r_mbytes_per_sec": 0, 00:05:46.599 "w_mbytes_per_sec": 0 00:05:46.599 }, 00:05:46.599 "claimed": false, 00:05:46.599 "zoned": false, 00:05:46.599 "supported_io_types": { 00:05:46.599 "read": true, 00:05:46.599 "write": true, 00:05:46.599 "unmap": true, 00:05:46.599 "flush": true, 00:05:46.599 "reset": true, 00:05:46.599 "nvme_admin": false, 00:05:46.599 "nvme_io": false, 00:05:46.599 "nvme_io_md": false, 00:05:46.599 "write_zeroes": true, 00:05:46.599 "zcopy": true, 00:05:46.599 "get_zone_info": false, 00:05:46.599 "zone_management": false, 00:05:46.599 "zone_append": false, 00:05:46.599 "compare": false, 00:05:46.599 "compare_and_write": false, 00:05:46.599 "abort": true, 00:05:46.599 "seek_hole": false, 00:05:46.599 "seek_data": false, 00:05:46.599 "copy": true, 00:05:46.599 "nvme_iov_md": false 00:05:46.599 }, 00:05:46.599 "memory_domains": [ 00:05:46.599 { 00:05:46.599 "dma_device_id": "system", 00:05:46.599 "dma_device_type": 1 00:05:46.599 }, 00:05:46.599 { 00:05:46.599 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:46.599 "dma_device_type": 2 00:05:46.599 } 00:05:46.599 ], 00:05:46.599 "driver_specific": {} 00:05:46.599 } 00:05:46.599 ]' 00:05:46.599 15:10:14 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # jq length 00:05:46.599 15:10:14 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # '[' 1 == 1 ']' 00:05:46.599 15:10:14 rpc.rpc_plugins -- rpc/rpc.sh@34 -- # rpc_cmd --plugin rpc_plugin delete_malloc Malloc1 00:05:46.599 15:10:14 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:46.599 15:10:14 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:05:46.599 15:10:14 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:46.599 15:10:14 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # rpc_cmd bdev_get_bdevs 00:05:46.599 15:10:14 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:46.599 15:10:14 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:05:46.599 15:10:14 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:46.599 15:10:14 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # bdevs='[]' 00:05:46.599 15:10:14 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # jq length 00:05:46.599 ************************************ 00:05:46.599 END TEST rpc_plugins 00:05:46.599 ************************************ 00:05:46.599 15:10:14 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # '[' 0 == 0 ']' 00:05:46.599 00:05:46.599 real 0m0.169s 00:05:46.599 user 0m0.099s 00:05:46.600 sys 0m0.028s 00:05:46.600 15:10:14 rpc.rpc_plugins -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:46.600 15:10:14 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:05:46.600 15:10:14 rpc -- rpc/rpc.sh@75 -- # run_test rpc_trace_cmd_test rpc_trace_cmd_test 00:05:46.600 15:10:14 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:46.600 15:10:14 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:46.600 15:10:14 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:46.863 ************************************ 00:05:46.863 START TEST rpc_trace_cmd_test 00:05:46.863 ************************************ 00:05:46.863 15:10:14 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1129 -- # rpc_trace_cmd_test 00:05:46.863 15:10:14 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@40 -- # local info 00:05:46.863 15:10:14 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # rpc_cmd trace_get_info 00:05:46.863 15:10:14 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:46.863 15:10:14 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:05:46.863 15:10:14 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:46.863 15:10:14 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # info='{ 00:05:46.863 "tpoint_shm_path": "/dev/shm/spdk_tgt_trace.pid69301", 00:05:46.863 "tpoint_group_mask": "0x8", 00:05:46.863 "iscsi_conn": { 00:05:46.863 "mask": "0x2", 00:05:46.863 "tpoint_mask": "0x0" 00:05:46.863 }, 00:05:46.863 "scsi": { 00:05:46.863 "mask": "0x4", 00:05:46.863 "tpoint_mask": "0x0" 00:05:46.863 }, 00:05:46.863 "bdev": { 00:05:46.863 "mask": "0x8", 00:05:46.864 "tpoint_mask": "0xffffffffffffffff" 00:05:46.864 }, 00:05:46.864 "nvmf_rdma": { 00:05:46.864 "mask": "0x10", 00:05:46.864 "tpoint_mask": "0x0" 00:05:46.864 }, 00:05:46.864 "nvmf_tcp": { 00:05:46.864 "mask": "0x20", 00:05:46.864 "tpoint_mask": "0x0" 00:05:46.864 }, 00:05:46.864 "ftl": { 00:05:46.864 "mask": "0x40", 00:05:46.864 "tpoint_mask": "0x0" 00:05:46.864 }, 00:05:46.864 "blobfs": { 00:05:46.864 "mask": "0x80", 00:05:46.864 "tpoint_mask": "0x0" 00:05:46.864 }, 00:05:46.864 "dsa": { 00:05:46.864 "mask": "0x200", 00:05:46.864 "tpoint_mask": "0x0" 00:05:46.864 }, 00:05:46.864 "thread": { 00:05:46.864 "mask": "0x400", 00:05:46.864 "tpoint_mask": "0x0" 00:05:46.864 }, 00:05:46.864 "nvme_pcie": { 00:05:46.864 "mask": "0x800", 00:05:46.864 "tpoint_mask": "0x0" 00:05:46.864 }, 00:05:46.864 "iaa": { 00:05:46.864 "mask": "0x1000", 00:05:46.864 "tpoint_mask": "0x0" 00:05:46.864 }, 00:05:46.864 "nvme_tcp": { 00:05:46.864 "mask": "0x2000", 00:05:46.864 "tpoint_mask": "0x0" 00:05:46.864 }, 00:05:46.864 "bdev_nvme": { 00:05:46.864 "mask": "0x4000", 00:05:46.864 "tpoint_mask": "0x0" 00:05:46.864 }, 00:05:46.864 "sock": { 00:05:46.864 "mask": "0x8000", 00:05:46.864 "tpoint_mask": "0x0" 00:05:46.864 }, 00:05:46.864 "blob": { 00:05:46.864 "mask": "0x10000", 00:05:46.864 "tpoint_mask": "0x0" 00:05:46.864 }, 00:05:46.864 "bdev_raid": { 00:05:46.864 "mask": "0x20000", 00:05:46.864 "tpoint_mask": "0x0" 00:05:46.864 }, 00:05:46.864 "scheduler": { 00:05:46.864 "mask": "0x40000", 00:05:46.864 "tpoint_mask": "0x0" 00:05:46.864 } 00:05:46.864 }' 00:05:46.864 15:10:14 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # jq length 00:05:46.864 15:10:14 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # '[' 19 -gt 2 ']' 00:05:46.864 15:10:14 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # jq 'has("tpoint_group_mask")' 00:05:46.864 15:10:14 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # '[' true = true ']' 00:05:46.864 15:10:14 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # jq 'has("tpoint_shm_path")' 00:05:46.864 15:10:14 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # '[' true = true ']' 00:05:46.864 15:10:14 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # jq 'has("bdev")' 00:05:46.864 15:10:14 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # '[' true = true ']' 00:05:46.864 15:10:14 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # jq -r .bdev.tpoint_mask 00:05:47.125 15:10:14 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # '[' 0xffffffffffffffff '!=' 0x0 ']' 00:05:47.125 ************************************ 00:05:47.125 END TEST rpc_trace_cmd_test 00:05:47.125 ************************************ 00:05:47.125 00:05:47.125 real 0m0.267s 00:05:47.125 user 0m0.205s 00:05:47.125 sys 0m0.039s 00:05:47.125 15:10:14 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:47.125 15:10:14 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:05:47.125 15:10:15 rpc -- rpc/rpc.sh@76 -- # [[ 0 -eq 1 ]] 00:05:47.125 15:10:15 rpc -- rpc/rpc.sh@80 -- # rpc=rpc_cmd 00:05:47.125 15:10:15 rpc -- rpc/rpc.sh@81 -- # run_test rpc_daemon_integrity rpc_integrity 00:05:47.125 15:10:15 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:47.125 15:10:15 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:47.125 15:10:15 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:47.125 ************************************ 00:05:47.125 START TEST rpc_daemon_integrity 00:05:47.125 ************************************ 00:05:47.125 15:10:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1129 -- # rpc_integrity 00:05:47.125 15:10:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:05:47.125 15:10:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:47.125 15:10:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:47.125 15:10:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:47.125 15:10:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:05:47.125 15:10:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # jq length 00:05:47.125 15:10:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:05:47.125 15:10:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:05:47.125 15:10:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:47.125 15:10:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:47.125 15:10:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:47.125 15:10:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc2 00:05:47.125 15:10:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:05:47.125 15:10:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:47.125 15:10:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:47.125 15:10:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:47.125 15:10:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:05:47.125 { 00:05:47.125 "name": "Malloc2", 00:05:47.125 "aliases": [ 00:05:47.125 "4729309c-6e39-4c98-b535-2372b5c2c9fc" 00:05:47.125 ], 00:05:47.125 "product_name": "Malloc disk", 00:05:47.125 "block_size": 512, 00:05:47.125 "num_blocks": 16384, 00:05:47.125 "uuid": "4729309c-6e39-4c98-b535-2372b5c2c9fc", 00:05:47.125 "assigned_rate_limits": { 00:05:47.125 "rw_ios_per_sec": 0, 00:05:47.125 "rw_mbytes_per_sec": 0, 00:05:47.125 "r_mbytes_per_sec": 0, 00:05:47.125 "w_mbytes_per_sec": 0 00:05:47.125 }, 00:05:47.125 "claimed": false, 00:05:47.125 "zoned": false, 00:05:47.125 "supported_io_types": { 00:05:47.125 "read": true, 00:05:47.125 "write": true, 00:05:47.125 "unmap": true, 00:05:47.125 "flush": true, 00:05:47.125 "reset": true, 00:05:47.125 "nvme_admin": false, 00:05:47.125 "nvme_io": false, 00:05:47.125 "nvme_io_md": false, 00:05:47.125 "write_zeroes": true, 00:05:47.125 "zcopy": true, 00:05:47.125 "get_zone_info": false, 00:05:47.125 "zone_management": false, 00:05:47.125 "zone_append": false, 00:05:47.125 "compare": false, 00:05:47.125 "compare_and_write": false, 00:05:47.125 "abort": true, 00:05:47.125 "seek_hole": false, 00:05:47.125 "seek_data": false, 00:05:47.125 "copy": true, 00:05:47.125 "nvme_iov_md": false 00:05:47.125 }, 00:05:47.125 "memory_domains": [ 00:05:47.125 { 00:05:47.125 "dma_device_id": "system", 00:05:47.125 "dma_device_type": 1 00:05:47.125 }, 00:05:47.125 { 00:05:47.125 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:47.125 "dma_device_type": 2 00:05:47.125 } 00:05:47.125 ], 00:05:47.125 "driver_specific": {} 00:05:47.125 } 00:05:47.125 ]' 00:05:47.125 15:10:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # jq length 00:05:47.125 15:10:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:05:47.125 15:10:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc2 -p Passthru0 00:05:47.125 15:10:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:47.125 15:10:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:47.125 [2024-11-27 15:10:15.175717] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on Malloc2 00:05:47.125 [2024-11-27 15:10:15.175809] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:05:47.125 [2024-11-27 15:10:15.175864] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:05:47.125 [2024-11-27 15:10:15.175875] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:05:47.125 [2024-11-27 15:10:15.178449] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:05:47.125 [2024-11-27 15:10:15.178509] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:05:47.125 Passthru0 00:05:47.125 15:10:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:47.126 15:10:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:05:47.126 15:10:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:47.126 15:10:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:47.126 15:10:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:47.126 15:10:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:05:47.126 { 00:05:47.126 "name": "Malloc2", 00:05:47.126 "aliases": [ 00:05:47.126 "4729309c-6e39-4c98-b535-2372b5c2c9fc" 00:05:47.126 ], 00:05:47.126 "product_name": "Malloc disk", 00:05:47.126 "block_size": 512, 00:05:47.126 "num_blocks": 16384, 00:05:47.126 "uuid": "4729309c-6e39-4c98-b535-2372b5c2c9fc", 00:05:47.126 "assigned_rate_limits": { 00:05:47.126 "rw_ios_per_sec": 0, 00:05:47.126 "rw_mbytes_per_sec": 0, 00:05:47.126 "r_mbytes_per_sec": 0, 00:05:47.126 "w_mbytes_per_sec": 0 00:05:47.126 }, 00:05:47.126 "claimed": true, 00:05:47.126 "claim_type": "exclusive_write", 00:05:47.126 "zoned": false, 00:05:47.126 "supported_io_types": { 00:05:47.126 "read": true, 00:05:47.126 "write": true, 00:05:47.126 "unmap": true, 00:05:47.126 "flush": true, 00:05:47.126 "reset": true, 00:05:47.126 "nvme_admin": false, 00:05:47.126 "nvme_io": false, 00:05:47.126 "nvme_io_md": false, 00:05:47.126 "write_zeroes": true, 00:05:47.126 "zcopy": true, 00:05:47.126 "get_zone_info": false, 00:05:47.126 "zone_management": false, 00:05:47.126 "zone_append": false, 00:05:47.126 "compare": false, 00:05:47.126 "compare_and_write": false, 00:05:47.126 "abort": true, 00:05:47.126 "seek_hole": false, 00:05:47.126 "seek_data": false, 00:05:47.126 "copy": true, 00:05:47.126 "nvme_iov_md": false 00:05:47.126 }, 00:05:47.126 "memory_domains": [ 00:05:47.126 { 00:05:47.126 "dma_device_id": "system", 00:05:47.126 "dma_device_type": 1 00:05:47.126 }, 00:05:47.126 { 00:05:47.126 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:47.126 "dma_device_type": 2 00:05:47.126 } 00:05:47.126 ], 00:05:47.126 "driver_specific": {} 00:05:47.126 }, 00:05:47.126 { 00:05:47.126 "name": "Passthru0", 00:05:47.126 "aliases": [ 00:05:47.126 "83ed6033-172a-531b-86df-a8f027ba5d2b" 00:05:47.126 ], 00:05:47.126 "product_name": "passthru", 00:05:47.126 "block_size": 512, 00:05:47.126 "num_blocks": 16384, 00:05:47.126 "uuid": "83ed6033-172a-531b-86df-a8f027ba5d2b", 00:05:47.126 "assigned_rate_limits": { 00:05:47.126 "rw_ios_per_sec": 0, 00:05:47.126 "rw_mbytes_per_sec": 0, 00:05:47.126 "r_mbytes_per_sec": 0, 00:05:47.126 "w_mbytes_per_sec": 0 00:05:47.126 }, 00:05:47.126 "claimed": false, 00:05:47.126 "zoned": false, 00:05:47.126 "supported_io_types": { 00:05:47.126 "read": true, 00:05:47.126 "write": true, 00:05:47.126 "unmap": true, 00:05:47.126 "flush": true, 00:05:47.126 "reset": true, 00:05:47.126 "nvme_admin": false, 00:05:47.126 "nvme_io": false, 00:05:47.126 "nvme_io_md": false, 00:05:47.126 "write_zeroes": true, 00:05:47.126 "zcopy": true, 00:05:47.126 "get_zone_info": false, 00:05:47.126 "zone_management": false, 00:05:47.126 "zone_append": false, 00:05:47.126 "compare": false, 00:05:47.126 "compare_and_write": false, 00:05:47.126 "abort": true, 00:05:47.126 "seek_hole": false, 00:05:47.126 "seek_data": false, 00:05:47.126 "copy": true, 00:05:47.126 "nvme_iov_md": false 00:05:47.126 }, 00:05:47.126 "memory_domains": [ 00:05:47.126 { 00:05:47.126 "dma_device_id": "system", 00:05:47.126 "dma_device_type": 1 00:05:47.126 }, 00:05:47.126 { 00:05:47.126 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:47.126 "dma_device_type": 2 00:05:47.126 } 00:05:47.126 ], 00:05:47.126 "driver_specific": { 00:05:47.126 "passthru": { 00:05:47.126 "name": "Passthru0", 00:05:47.126 "base_bdev_name": "Malloc2" 00:05:47.126 } 00:05:47.126 } 00:05:47.126 } 00:05:47.126 ]' 00:05:47.126 15:10:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # jq length 00:05:47.386 15:10:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:05:47.386 15:10:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:05:47.386 15:10:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:47.386 15:10:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:47.386 15:10:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:47.386 15:10:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc2 00:05:47.386 15:10:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:47.386 15:10:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:47.386 15:10:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:47.386 15:10:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:05:47.386 15:10:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:47.386 15:10:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:47.386 15:10:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:47.386 15:10:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:05:47.386 15:10:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # jq length 00:05:47.386 15:10:15 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:05:47.386 00:05:47.386 real 0m0.303s 00:05:47.386 user 0m0.184s 00:05:47.386 sys 0m0.052s 00:05:47.386 15:10:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:47.386 ************************************ 00:05:47.386 END TEST rpc_daemon_integrity 00:05:47.386 ************************************ 00:05:47.386 15:10:15 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:47.386 15:10:15 rpc -- rpc/rpc.sh@83 -- # trap - SIGINT SIGTERM EXIT 00:05:47.386 15:10:15 rpc -- rpc/rpc.sh@84 -- # killprocess 69301 00:05:47.387 15:10:15 rpc -- common/autotest_common.sh@954 -- # '[' -z 69301 ']' 00:05:47.387 15:10:15 rpc -- common/autotest_common.sh@958 -- # kill -0 69301 00:05:47.387 15:10:15 rpc -- common/autotest_common.sh@959 -- # uname 00:05:47.387 15:10:15 rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:47.387 15:10:15 rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69301 00:05:47.387 15:10:15 rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:47.387 15:10:15 rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:47.387 killing process with pid 69301 00:05:47.387 15:10:15 rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69301' 00:05:47.387 15:10:15 rpc -- common/autotest_common.sh@973 -- # kill 69301 00:05:47.387 15:10:15 rpc -- common/autotest_common.sh@978 -- # wait 69301 00:05:47.957 00:05:47.957 real 0m2.919s 00:05:47.957 user 0m3.530s 00:05:47.957 sys 0m0.885s 00:05:47.957 15:10:15 rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:47.957 15:10:15 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:47.957 ************************************ 00:05:47.957 END TEST rpc 00:05:47.957 ************************************ 00:05:47.957 15:10:15 -- spdk/autotest.sh@157 -- # run_test skip_rpc /home/vagrant/spdk_repo/spdk/test/rpc/skip_rpc.sh 00:05:47.957 15:10:15 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:47.957 15:10:15 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:47.957 15:10:15 -- common/autotest_common.sh@10 -- # set +x 00:05:47.957 ************************************ 00:05:47.957 START TEST skip_rpc 00:05:47.957 ************************************ 00:05:47.957 15:10:15 skip_rpc -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/rpc/skip_rpc.sh 00:05:47.957 * Looking for test storage... 00:05:47.957 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc 00:05:47.957 15:10:15 skip_rpc -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:05:47.957 15:10:15 skip_rpc -- common/autotest_common.sh@1693 -- # lcov --version 00:05:47.957 15:10:15 skip_rpc -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:05:48.218 15:10:16 skip_rpc -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:05:48.218 15:10:16 skip_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:48.218 15:10:16 skip_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:48.218 15:10:16 skip_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:48.218 15:10:16 skip_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:05:48.218 15:10:16 skip_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:05:48.218 15:10:16 skip_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:05:48.218 15:10:16 skip_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:05:48.218 15:10:16 skip_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:05:48.218 15:10:16 skip_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:05:48.218 15:10:16 skip_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:05:48.218 15:10:16 skip_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:48.218 15:10:16 skip_rpc -- scripts/common.sh@344 -- # case "$op" in 00:05:48.218 15:10:16 skip_rpc -- scripts/common.sh@345 -- # : 1 00:05:48.218 15:10:16 skip_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:48.218 15:10:16 skip_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:48.218 15:10:16 skip_rpc -- scripts/common.sh@365 -- # decimal 1 00:05:48.218 15:10:16 skip_rpc -- scripts/common.sh@353 -- # local d=1 00:05:48.218 15:10:16 skip_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:48.218 15:10:16 skip_rpc -- scripts/common.sh@355 -- # echo 1 00:05:48.218 15:10:16 skip_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:05:48.218 15:10:16 skip_rpc -- scripts/common.sh@366 -- # decimal 2 00:05:48.218 15:10:16 skip_rpc -- scripts/common.sh@353 -- # local d=2 00:05:48.218 15:10:16 skip_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:48.218 15:10:16 skip_rpc -- scripts/common.sh@355 -- # echo 2 00:05:48.218 15:10:16 skip_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:05:48.218 15:10:16 skip_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:48.218 15:10:16 skip_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:48.218 15:10:16 skip_rpc -- scripts/common.sh@368 -- # return 0 00:05:48.218 15:10:16 skip_rpc -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:48.218 15:10:16 skip_rpc -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:05:48.218 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:48.218 --rc genhtml_branch_coverage=1 00:05:48.218 --rc genhtml_function_coverage=1 00:05:48.218 --rc genhtml_legend=1 00:05:48.218 --rc geninfo_all_blocks=1 00:05:48.218 --rc geninfo_unexecuted_blocks=1 00:05:48.218 00:05:48.218 ' 00:05:48.218 15:10:16 skip_rpc -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:05:48.218 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:48.218 --rc genhtml_branch_coverage=1 00:05:48.218 --rc genhtml_function_coverage=1 00:05:48.218 --rc genhtml_legend=1 00:05:48.218 --rc geninfo_all_blocks=1 00:05:48.218 --rc geninfo_unexecuted_blocks=1 00:05:48.218 00:05:48.218 ' 00:05:48.218 15:10:16 skip_rpc -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:05:48.218 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:48.218 --rc genhtml_branch_coverage=1 00:05:48.218 --rc genhtml_function_coverage=1 00:05:48.218 --rc genhtml_legend=1 00:05:48.218 --rc geninfo_all_blocks=1 00:05:48.218 --rc geninfo_unexecuted_blocks=1 00:05:48.218 00:05:48.218 ' 00:05:48.218 15:10:16 skip_rpc -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:05:48.218 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:48.218 --rc genhtml_branch_coverage=1 00:05:48.218 --rc genhtml_function_coverage=1 00:05:48.218 --rc genhtml_legend=1 00:05:48.218 --rc geninfo_all_blocks=1 00:05:48.218 --rc geninfo_unexecuted_blocks=1 00:05:48.218 00:05:48.218 ' 00:05:48.218 15:10:16 skip_rpc -- rpc/skip_rpc.sh@11 -- # CONFIG_PATH=/home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:05:48.218 15:10:16 skip_rpc -- rpc/skip_rpc.sh@12 -- # LOG_PATH=/home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:05:48.218 15:10:16 skip_rpc -- rpc/skip_rpc.sh@73 -- # run_test skip_rpc test_skip_rpc 00:05:48.218 15:10:16 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:48.218 15:10:16 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:48.218 15:10:16 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:48.218 ************************************ 00:05:48.218 START TEST skip_rpc 00:05:48.218 ************************************ 00:05:48.218 15:10:16 skip_rpc.skip_rpc -- common/autotest_common.sh@1129 -- # test_skip_rpc 00:05:48.219 15:10:16 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@16 -- # local spdk_pid=69508 00:05:48.219 15:10:16 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@15 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 00:05:48.219 15:10:16 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@18 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:05:48.219 15:10:16 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@19 -- # sleep 5 00:05:48.219 [2024-11-27 15:10:16.198076] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:05:48.219 [2024-11-27 15:10:16.198283] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69508 ] 00:05:48.479 [2024-11-27 15:10:16.366859] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:48.479 [2024-11-27 15:10:16.397225] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:53.764 15:10:21 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@21 -- # NOT rpc_cmd spdk_get_version 00:05:53.764 15:10:21 skip_rpc.skip_rpc -- common/autotest_common.sh@652 -- # local es=0 00:05:53.764 15:10:21 skip_rpc.skip_rpc -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd spdk_get_version 00:05:53.764 15:10:21 skip_rpc.skip_rpc -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:05:53.764 15:10:21 skip_rpc.skip_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:53.764 15:10:21 skip_rpc.skip_rpc -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:05:53.764 15:10:21 skip_rpc.skip_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:53.764 15:10:21 skip_rpc.skip_rpc -- common/autotest_common.sh@655 -- # rpc_cmd spdk_get_version 00:05:53.764 15:10:21 skip_rpc.skip_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:53.764 15:10:21 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:53.764 15:10:21 skip_rpc.skip_rpc -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:05:53.764 15:10:21 skip_rpc.skip_rpc -- common/autotest_common.sh@655 -- # es=1 00:05:53.764 15:10:21 skip_rpc.skip_rpc -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:05:53.764 15:10:21 skip_rpc.skip_rpc -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:05:53.764 15:10:21 skip_rpc.skip_rpc -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:05:53.764 15:10:21 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@22 -- # trap - SIGINT SIGTERM EXIT 00:05:53.764 15:10:21 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@23 -- # killprocess 69508 00:05:53.764 15:10:21 skip_rpc.skip_rpc -- common/autotest_common.sh@954 -- # '[' -z 69508 ']' 00:05:53.764 15:10:21 skip_rpc.skip_rpc -- common/autotest_common.sh@958 -- # kill -0 69508 00:05:53.764 15:10:21 skip_rpc.skip_rpc -- common/autotest_common.sh@959 -- # uname 00:05:53.764 15:10:21 skip_rpc.skip_rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:53.764 15:10:21 skip_rpc.skip_rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69508 00:05:53.764 15:10:21 skip_rpc.skip_rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:53.764 15:10:21 skip_rpc.skip_rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:53.764 15:10:21 skip_rpc.skip_rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69508' 00:05:53.764 killing process with pid 69508 00:05:53.764 15:10:21 skip_rpc.skip_rpc -- common/autotest_common.sh@973 -- # kill 69508 00:05:53.764 15:10:21 skip_rpc.skip_rpc -- common/autotest_common.sh@978 -- # wait 69508 00:05:53.764 00:05:53.764 real 0m5.421s 00:05:53.764 user 0m5.018s 00:05:53.764 sys 0m0.317s 00:05:53.764 15:10:21 skip_rpc.skip_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:53.764 ************************************ 00:05:53.764 END TEST skip_rpc 00:05:53.764 ************************************ 00:05:53.764 15:10:21 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:53.764 15:10:21 skip_rpc -- rpc/skip_rpc.sh@74 -- # run_test skip_rpc_with_json test_skip_rpc_with_json 00:05:53.764 15:10:21 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:53.764 15:10:21 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:53.764 15:10:21 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:53.764 ************************************ 00:05:53.764 START TEST skip_rpc_with_json 00:05:53.764 ************************************ 00:05:53.764 15:10:21 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1129 -- # test_skip_rpc_with_json 00:05:53.764 15:10:21 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@44 -- # gen_json_config 00:05:53.764 15:10:21 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@28 -- # local spdk_pid=69596 00:05:53.764 15:10:21 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@27 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:05:53.764 15:10:21 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@30 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:05:53.764 15:10:21 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@31 -- # waitforlisten 69596 00:05:53.764 15:10:21 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@835 -- # '[' -z 69596 ']' 00:05:53.764 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:53.764 15:10:21 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:53.764 15:10:21 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:53.764 15:10:21 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:53.764 15:10:21 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:53.764 15:10:21 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:05:53.764 [2024-11-27 15:10:21.684329] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:05:53.764 [2024-11-27 15:10:21.684560] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69596 ] 00:05:53.764 [2024-11-27 15:10:21.845147] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:54.024 [2024-11-27 15:10:21.875083] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:54.594 15:10:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:54.594 15:10:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@868 -- # return 0 00:05:54.594 15:10:22 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_get_transports --trtype tcp 00:05:54.594 15:10:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:54.594 15:10:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:05:54.594 [2024-11-27 15:10:22.519145] nvmf_rpc.c:2706:rpc_nvmf_get_transports: *ERROR*: transport 'tcp' does not exist 00:05:54.594 request: 00:05:54.594 { 00:05:54.594 "trtype": "tcp", 00:05:54.594 "method": "nvmf_get_transports", 00:05:54.594 "req_id": 1 00:05:54.594 } 00:05:54.594 Got JSON-RPC error response 00:05:54.594 response: 00:05:54.594 { 00:05:54.594 "code": -19, 00:05:54.594 "message": "No such device" 00:05:54.594 } 00:05:54.594 15:10:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:05:54.594 15:10:22 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_create_transport -t tcp 00:05:54.594 15:10:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:54.594 15:10:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:05:54.594 [2024-11-27 15:10:22.531237] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:05:54.594 15:10:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:54.594 15:10:22 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@36 -- # rpc_cmd save_config 00:05:54.594 15:10:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:54.594 15:10:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:05:54.854 15:10:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:54.854 15:10:22 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@37 -- # cat /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:05:54.854 { 00:05:54.854 "subsystems": [ 00:05:54.854 { 00:05:54.854 "subsystem": "fsdev", 00:05:54.854 "config": [ 00:05:54.854 { 00:05:54.854 "method": "fsdev_set_opts", 00:05:54.854 "params": { 00:05:54.854 "fsdev_io_pool_size": 65535, 00:05:54.854 "fsdev_io_cache_size": 256 00:05:54.854 } 00:05:54.854 } 00:05:54.854 ] 00:05:54.854 }, 00:05:54.854 { 00:05:54.854 "subsystem": "keyring", 00:05:54.854 "config": [] 00:05:54.854 }, 00:05:54.854 { 00:05:54.854 "subsystem": "iobuf", 00:05:54.854 "config": [ 00:05:54.854 { 00:05:54.854 "method": "iobuf_set_options", 00:05:54.854 "params": { 00:05:54.854 "small_pool_count": 8192, 00:05:54.854 "large_pool_count": 1024, 00:05:54.854 "small_bufsize": 8192, 00:05:54.854 "large_bufsize": 135168, 00:05:54.854 "enable_numa": false 00:05:54.854 } 00:05:54.854 } 00:05:54.854 ] 00:05:54.854 }, 00:05:54.854 { 00:05:54.854 "subsystem": "sock", 00:05:54.854 "config": [ 00:05:54.854 { 00:05:54.854 "method": "sock_set_default_impl", 00:05:54.854 "params": { 00:05:54.854 "impl_name": "posix" 00:05:54.854 } 00:05:54.854 }, 00:05:54.854 { 00:05:54.854 "method": "sock_impl_set_options", 00:05:54.854 "params": { 00:05:54.854 "impl_name": "ssl", 00:05:54.854 "recv_buf_size": 4096, 00:05:54.854 "send_buf_size": 4096, 00:05:54.854 "enable_recv_pipe": true, 00:05:54.854 "enable_quickack": false, 00:05:54.854 "enable_placement_id": 0, 00:05:54.854 "enable_zerocopy_send_server": true, 00:05:54.854 "enable_zerocopy_send_client": false, 00:05:54.854 "zerocopy_threshold": 0, 00:05:54.854 "tls_version": 0, 00:05:54.854 "enable_ktls": false 00:05:54.854 } 00:05:54.854 }, 00:05:54.854 { 00:05:54.854 "method": "sock_impl_set_options", 00:05:54.854 "params": { 00:05:54.854 "impl_name": "posix", 00:05:54.854 "recv_buf_size": 2097152, 00:05:54.854 "send_buf_size": 2097152, 00:05:54.854 "enable_recv_pipe": true, 00:05:54.854 "enable_quickack": false, 00:05:54.854 "enable_placement_id": 0, 00:05:54.854 "enable_zerocopy_send_server": true, 00:05:54.854 "enable_zerocopy_send_client": false, 00:05:54.854 "zerocopy_threshold": 0, 00:05:54.854 "tls_version": 0, 00:05:54.854 "enable_ktls": false 00:05:54.854 } 00:05:54.854 } 00:05:54.854 ] 00:05:54.854 }, 00:05:54.854 { 00:05:54.854 "subsystem": "vmd", 00:05:54.854 "config": [] 00:05:54.854 }, 00:05:54.854 { 00:05:54.854 "subsystem": "accel", 00:05:54.854 "config": [ 00:05:54.854 { 00:05:54.854 "method": "accel_set_options", 00:05:54.854 "params": { 00:05:54.854 "small_cache_size": 128, 00:05:54.854 "large_cache_size": 16, 00:05:54.854 "task_count": 2048, 00:05:54.854 "sequence_count": 2048, 00:05:54.854 "buf_count": 2048 00:05:54.854 } 00:05:54.854 } 00:05:54.854 ] 00:05:54.854 }, 00:05:54.854 { 00:05:54.854 "subsystem": "bdev", 00:05:54.854 "config": [ 00:05:54.854 { 00:05:54.854 "method": "bdev_set_options", 00:05:54.854 "params": { 00:05:54.854 "bdev_io_pool_size": 65535, 00:05:54.854 "bdev_io_cache_size": 256, 00:05:54.854 "bdev_auto_examine": true, 00:05:54.854 "iobuf_small_cache_size": 128, 00:05:54.854 "iobuf_large_cache_size": 16 00:05:54.854 } 00:05:54.854 }, 00:05:54.854 { 00:05:54.854 "method": "bdev_raid_set_options", 00:05:54.854 "params": { 00:05:54.854 "process_window_size_kb": 1024, 00:05:54.854 "process_max_bandwidth_mb_sec": 0 00:05:54.854 } 00:05:54.854 }, 00:05:54.854 { 00:05:54.854 "method": "bdev_iscsi_set_options", 00:05:54.854 "params": { 00:05:54.854 "timeout_sec": 30 00:05:54.854 } 00:05:54.854 }, 00:05:54.854 { 00:05:54.854 "method": "bdev_nvme_set_options", 00:05:54.854 "params": { 00:05:54.854 "action_on_timeout": "none", 00:05:54.854 "timeout_us": 0, 00:05:54.854 "timeout_admin_us": 0, 00:05:54.854 "keep_alive_timeout_ms": 10000, 00:05:54.854 "arbitration_burst": 0, 00:05:54.854 "low_priority_weight": 0, 00:05:54.854 "medium_priority_weight": 0, 00:05:54.854 "high_priority_weight": 0, 00:05:54.854 "nvme_adminq_poll_period_us": 10000, 00:05:54.854 "nvme_ioq_poll_period_us": 0, 00:05:54.854 "io_queue_requests": 0, 00:05:54.854 "delay_cmd_submit": true, 00:05:54.854 "transport_retry_count": 4, 00:05:54.854 "bdev_retry_count": 3, 00:05:54.854 "transport_ack_timeout": 0, 00:05:54.854 "ctrlr_loss_timeout_sec": 0, 00:05:54.854 "reconnect_delay_sec": 0, 00:05:54.854 "fast_io_fail_timeout_sec": 0, 00:05:54.854 "disable_auto_failback": false, 00:05:54.854 "generate_uuids": false, 00:05:54.854 "transport_tos": 0, 00:05:54.854 "nvme_error_stat": false, 00:05:54.854 "rdma_srq_size": 0, 00:05:54.854 "io_path_stat": false, 00:05:54.854 "allow_accel_sequence": false, 00:05:54.854 "rdma_max_cq_size": 0, 00:05:54.854 "rdma_cm_event_timeout_ms": 0, 00:05:54.854 "dhchap_digests": [ 00:05:54.854 "sha256", 00:05:54.854 "sha384", 00:05:54.854 "sha512" 00:05:54.854 ], 00:05:54.854 "dhchap_dhgroups": [ 00:05:54.854 "null", 00:05:54.854 "ffdhe2048", 00:05:54.854 "ffdhe3072", 00:05:54.854 "ffdhe4096", 00:05:54.854 "ffdhe6144", 00:05:54.854 "ffdhe8192" 00:05:54.854 ] 00:05:54.854 } 00:05:54.854 }, 00:05:54.854 { 00:05:54.854 "method": "bdev_nvme_set_hotplug", 00:05:54.854 "params": { 00:05:54.854 "period_us": 100000, 00:05:54.854 "enable": false 00:05:54.854 } 00:05:54.854 }, 00:05:54.854 { 00:05:54.854 "method": "bdev_wait_for_examine" 00:05:54.854 } 00:05:54.854 ] 00:05:54.854 }, 00:05:54.854 { 00:05:54.854 "subsystem": "scsi", 00:05:54.854 "config": null 00:05:54.854 }, 00:05:54.854 { 00:05:54.854 "subsystem": "scheduler", 00:05:54.854 "config": [ 00:05:54.854 { 00:05:54.854 "method": "framework_set_scheduler", 00:05:54.854 "params": { 00:05:54.854 "name": "static" 00:05:54.854 } 00:05:54.854 } 00:05:54.854 ] 00:05:54.854 }, 00:05:54.854 { 00:05:54.854 "subsystem": "vhost_scsi", 00:05:54.854 "config": [] 00:05:54.854 }, 00:05:54.854 { 00:05:54.854 "subsystem": "vhost_blk", 00:05:54.854 "config": [] 00:05:54.854 }, 00:05:54.854 { 00:05:54.854 "subsystem": "ublk", 00:05:54.854 "config": [] 00:05:54.854 }, 00:05:54.854 { 00:05:54.854 "subsystem": "nbd", 00:05:54.854 "config": [] 00:05:54.854 }, 00:05:54.854 { 00:05:54.854 "subsystem": "nvmf", 00:05:54.854 "config": [ 00:05:54.854 { 00:05:54.854 "method": "nvmf_set_config", 00:05:54.854 "params": { 00:05:54.854 "discovery_filter": "match_any", 00:05:54.854 "admin_cmd_passthru": { 00:05:54.854 "identify_ctrlr": false 00:05:54.854 }, 00:05:54.854 "dhchap_digests": [ 00:05:54.854 "sha256", 00:05:54.854 "sha384", 00:05:54.854 "sha512" 00:05:54.854 ], 00:05:54.854 "dhchap_dhgroups": [ 00:05:54.854 "null", 00:05:54.854 "ffdhe2048", 00:05:54.854 "ffdhe3072", 00:05:54.854 "ffdhe4096", 00:05:54.854 "ffdhe6144", 00:05:54.854 "ffdhe8192" 00:05:54.854 ] 00:05:54.854 } 00:05:54.854 }, 00:05:54.854 { 00:05:54.855 "method": "nvmf_set_max_subsystems", 00:05:54.855 "params": { 00:05:54.855 "max_subsystems": 1024 00:05:54.855 } 00:05:54.855 }, 00:05:54.855 { 00:05:54.855 "method": "nvmf_set_crdt", 00:05:54.855 "params": { 00:05:54.855 "crdt1": 0, 00:05:54.855 "crdt2": 0, 00:05:54.855 "crdt3": 0 00:05:54.855 } 00:05:54.855 }, 00:05:54.855 { 00:05:54.855 "method": "nvmf_create_transport", 00:05:54.855 "params": { 00:05:54.855 "trtype": "TCP", 00:05:54.855 "max_queue_depth": 128, 00:05:54.855 "max_io_qpairs_per_ctrlr": 127, 00:05:54.855 "in_capsule_data_size": 4096, 00:05:54.855 "max_io_size": 131072, 00:05:54.855 "io_unit_size": 131072, 00:05:54.855 "max_aq_depth": 128, 00:05:54.855 "num_shared_buffers": 511, 00:05:54.855 "buf_cache_size": 4294967295, 00:05:54.855 "dif_insert_or_strip": false, 00:05:54.855 "zcopy": false, 00:05:54.855 "c2h_success": true, 00:05:54.855 "sock_priority": 0, 00:05:54.855 "abort_timeout_sec": 1, 00:05:54.855 "ack_timeout": 0, 00:05:54.855 "data_wr_pool_size": 0 00:05:54.855 } 00:05:54.855 } 00:05:54.855 ] 00:05:54.855 }, 00:05:54.855 { 00:05:54.855 "subsystem": "iscsi", 00:05:54.855 "config": [ 00:05:54.855 { 00:05:54.855 "method": "iscsi_set_options", 00:05:54.855 "params": { 00:05:54.855 "node_base": "iqn.2016-06.io.spdk", 00:05:54.855 "max_sessions": 128, 00:05:54.855 "max_connections_per_session": 2, 00:05:54.855 "max_queue_depth": 64, 00:05:54.855 "default_time2wait": 2, 00:05:54.855 "default_time2retain": 20, 00:05:54.855 "first_burst_length": 8192, 00:05:54.855 "immediate_data": true, 00:05:54.855 "allow_duplicated_isid": false, 00:05:54.855 "error_recovery_level": 0, 00:05:54.855 "nop_timeout": 60, 00:05:54.855 "nop_in_interval": 30, 00:05:54.855 "disable_chap": false, 00:05:54.855 "require_chap": false, 00:05:54.855 "mutual_chap": false, 00:05:54.855 "chap_group": 0, 00:05:54.855 "max_large_datain_per_connection": 64, 00:05:54.855 "max_r2t_per_connection": 4, 00:05:54.855 "pdu_pool_size": 36864, 00:05:54.855 "immediate_data_pool_size": 16384, 00:05:54.855 "data_out_pool_size": 2048 00:05:54.855 } 00:05:54.855 } 00:05:54.855 ] 00:05:54.855 } 00:05:54.855 ] 00:05:54.855 } 00:05:54.855 15:10:22 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@39 -- # trap - SIGINT SIGTERM EXIT 00:05:54.855 15:10:22 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@40 -- # killprocess 69596 00:05:54.855 15:10:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # '[' -z 69596 ']' 00:05:54.855 15:10:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@958 -- # kill -0 69596 00:05:54.855 15:10:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # uname 00:05:54.855 15:10:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:54.855 15:10:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69596 00:05:54.855 15:10:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:54.855 15:10:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:54.855 killing process with pid 69596 00:05:54.855 15:10:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69596' 00:05:54.855 15:10:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@973 -- # kill 69596 00:05:54.855 15:10:22 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@978 -- # wait 69596 00:05:55.115 15:10:23 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@47 -- # local spdk_pid=69619 00:05:55.115 15:10:23 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@46 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --json /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:05:55.115 15:10:23 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@48 -- # sleep 5 00:06:00.394 15:10:28 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@50 -- # killprocess 69619 00:06:00.394 15:10:28 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # '[' -z 69619 ']' 00:06:00.394 15:10:28 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@958 -- # kill -0 69619 00:06:00.394 15:10:28 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # uname 00:06:00.394 15:10:28 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:00.394 15:10:28 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69619 00:06:00.394 15:10:28 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:00.394 killing process with pid 69619 00:06:00.394 15:10:28 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:00.394 15:10:28 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69619' 00:06:00.394 15:10:28 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@973 -- # kill 69619 00:06:00.394 15:10:28 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@978 -- # wait 69619 00:06:00.654 15:10:28 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@51 -- # grep -q 'TCP Transport Init' /home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:06:00.654 15:10:28 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@52 -- # rm /home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:06:00.654 ************************************ 00:06:00.654 END TEST skip_rpc_with_json 00:06:00.654 ************************************ 00:06:00.654 00:06:00.654 real 0m6.976s 00:06:00.654 user 0m6.544s 00:06:00.654 sys 0m0.735s 00:06:00.654 15:10:28 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:00.654 15:10:28 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:06:00.654 15:10:28 skip_rpc -- rpc/skip_rpc.sh@75 -- # run_test skip_rpc_with_delay test_skip_rpc_with_delay 00:06:00.654 15:10:28 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:00.654 15:10:28 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:00.654 15:10:28 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:00.654 ************************************ 00:06:00.654 START TEST skip_rpc_with_delay 00:06:00.654 ************************************ 00:06:00.654 15:10:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1129 -- # test_skip_rpc_with_delay 00:06:00.654 15:10:28 skip_rpc.skip_rpc_with_delay -- rpc/skip_rpc.sh@57 -- # NOT /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:06:00.654 15:10:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@652 -- # local es=0 00:06:00.654 15:10:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@654 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:06:00.654 15:10:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@640 -- # local arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:00.654 15:10:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:00.654 15:10:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # type -t /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:00.654 15:10:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:00.654 15:10:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@646 -- # type -P /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:00.654 15:10:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:00.654 15:10:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@646 -- # arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:00.654 15:10:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@646 -- # [[ -x /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt ]] 00:06:00.654 15:10:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@655 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:06:00.654 [2024-11-27 15:10:28.729045] app.c: 842:spdk_app_start: *ERROR*: Cannot use '--wait-for-rpc' if no RPC server is going to be started. 00:06:00.913 15:10:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@655 -- # es=1 00:06:00.913 15:10:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:06:00.913 15:10:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:06:00.913 15:10:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:06:00.913 00:06:00.913 real 0m0.174s 00:06:00.913 user 0m0.092s 00:06:00.913 sys 0m0.080s 00:06:00.913 15:10:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:00.913 ************************************ 00:06:00.913 END TEST skip_rpc_with_delay 00:06:00.913 ************************************ 00:06:00.913 15:10:28 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@10 -- # set +x 00:06:00.913 15:10:28 skip_rpc -- rpc/skip_rpc.sh@77 -- # uname 00:06:00.913 15:10:28 skip_rpc -- rpc/skip_rpc.sh@77 -- # '[' Linux '!=' FreeBSD ']' 00:06:00.913 15:10:28 skip_rpc -- rpc/skip_rpc.sh@78 -- # run_test exit_on_failed_rpc_init test_exit_on_failed_rpc_init 00:06:00.913 15:10:28 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:00.913 15:10:28 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:00.913 15:10:28 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:00.913 ************************************ 00:06:00.913 START TEST exit_on_failed_rpc_init 00:06:00.913 ************************************ 00:06:00.913 15:10:28 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1129 -- # test_exit_on_failed_rpc_init 00:06:00.913 15:10:28 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@62 -- # local spdk_pid=69736 00:06:00.913 15:10:28 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@61 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:00.913 15:10:28 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@63 -- # waitforlisten 69736 00:06:00.913 15:10:28 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@835 -- # '[' -z 69736 ']' 00:06:00.913 15:10:28 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:00.913 15:10:28 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:00.913 15:10:28 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:00.913 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:00.913 15:10:28 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:00.913 15:10:28 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:06:00.913 [2024-11-27 15:10:28.968831] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:06:00.913 [2024-11-27 15:10:28.969095] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69736 ] 00:06:01.172 [2024-11-27 15:10:29.129289] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:01.172 [2024-11-27 15:10:29.159095] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:01.739 15:10:29 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:01.739 15:10:29 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@868 -- # return 0 00:06:01.739 15:10:29 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@65 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:06:01.739 15:10:29 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@67 -- # NOT /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:06:01.739 15:10:29 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@652 -- # local es=0 00:06:01.739 15:10:29 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@654 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:06:01.739 15:10:29 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@640 -- # local arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:01.739 15:10:29 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:01.739 15:10:29 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # type -t /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:01.739 15:10:29 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:01.739 15:10:29 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@646 -- # type -P /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:01.739 15:10:29 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:01.739 15:10:29 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@646 -- # arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:01.739 15:10:29 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@646 -- # [[ -x /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt ]] 00:06:01.739 15:10:29 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@655 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:06:01.999 [2024-11-27 15:10:29.909629] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:06:01.999 [2024-11-27 15:10:29.909769] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69754 ] 00:06:01.999 [2024-11-27 15:10:30.085392] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:02.259 [2024-11-27 15:10:30.115682] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:02.259 [2024-11-27 15:10:30.115912] rpc.c: 180:_spdk_rpc_listen: *ERROR*: RPC Unix domain socket path /var/tmp/spdk.sock in use. Specify another. 00:06:02.259 [2024-11-27 15:10:30.115958] rpc.c: 166:spdk_rpc_initialize: *ERROR*: Unable to start RPC service at /var/tmp/spdk.sock 00:06:02.259 [2024-11-27 15:10:30.115969] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:06:02.259 15:10:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@655 -- # es=234 00:06:02.259 15:10:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:06:02.259 15:10:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@664 -- # es=106 00:06:02.259 15:10:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@665 -- # case "$es" in 00:06:02.259 15:10:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@672 -- # es=1 00:06:02.259 15:10:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:06:02.259 15:10:30 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@69 -- # trap - SIGINT SIGTERM EXIT 00:06:02.259 15:10:30 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@70 -- # killprocess 69736 00:06:02.259 15:10:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@954 -- # '[' -z 69736 ']' 00:06:02.260 15:10:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@958 -- # kill -0 69736 00:06:02.260 15:10:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@959 -- # uname 00:06:02.260 15:10:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:02.260 15:10:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69736 00:06:02.260 15:10:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:02.260 15:10:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:02.260 15:10:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69736' 00:06:02.260 killing process with pid 69736 00:06:02.260 15:10:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@973 -- # kill 69736 00:06:02.260 15:10:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@978 -- # wait 69736 00:06:02.828 00:06:02.828 real 0m1.768s 00:06:02.828 user 0m1.922s 00:06:02.828 sys 0m0.510s 00:06:02.828 15:10:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:02.828 15:10:30 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:06:02.828 ************************************ 00:06:02.828 END TEST exit_on_failed_rpc_init 00:06:02.828 ************************************ 00:06:02.828 15:10:30 skip_rpc -- rpc/skip_rpc.sh@81 -- # rm /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:06:02.828 00:06:02.828 real 0m14.831s 00:06:02.828 user 0m13.796s 00:06:02.828 sys 0m1.926s 00:06:02.828 15:10:30 skip_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:02.828 15:10:30 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:02.828 ************************************ 00:06:02.828 END TEST skip_rpc 00:06:02.828 ************************************ 00:06:02.828 15:10:30 -- spdk/autotest.sh@158 -- # run_test rpc_client /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client.sh 00:06:02.828 15:10:30 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:02.828 15:10:30 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:02.828 15:10:30 -- common/autotest_common.sh@10 -- # set +x 00:06:02.828 ************************************ 00:06:02.828 START TEST rpc_client 00:06:02.828 ************************************ 00:06:02.828 15:10:30 rpc_client -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client.sh 00:06:02.828 * Looking for test storage... 00:06:02.828 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc_client 00:06:02.828 15:10:30 rpc_client -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:06:02.828 15:10:30 rpc_client -- common/autotest_common.sh@1693 -- # lcov --version 00:06:02.828 15:10:30 rpc_client -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:06:03.088 15:10:30 rpc_client -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:06:03.088 15:10:30 rpc_client -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:03.088 15:10:30 rpc_client -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:03.088 15:10:30 rpc_client -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:03.088 15:10:30 rpc_client -- scripts/common.sh@336 -- # IFS=.-: 00:06:03.088 15:10:30 rpc_client -- scripts/common.sh@336 -- # read -ra ver1 00:06:03.088 15:10:30 rpc_client -- scripts/common.sh@337 -- # IFS=.-: 00:06:03.088 15:10:30 rpc_client -- scripts/common.sh@337 -- # read -ra ver2 00:06:03.088 15:10:30 rpc_client -- scripts/common.sh@338 -- # local 'op=<' 00:06:03.088 15:10:30 rpc_client -- scripts/common.sh@340 -- # ver1_l=2 00:06:03.088 15:10:30 rpc_client -- scripts/common.sh@341 -- # ver2_l=1 00:06:03.088 15:10:30 rpc_client -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:03.088 15:10:30 rpc_client -- scripts/common.sh@344 -- # case "$op" in 00:06:03.088 15:10:30 rpc_client -- scripts/common.sh@345 -- # : 1 00:06:03.088 15:10:30 rpc_client -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:03.088 15:10:30 rpc_client -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:03.088 15:10:30 rpc_client -- scripts/common.sh@365 -- # decimal 1 00:06:03.088 15:10:30 rpc_client -- scripts/common.sh@353 -- # local d=1 00:06:03.088 15:10:30 rpc_client -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:03.088 15:10:30 rpc_client -- scripts/common.sh@355 -- # echo 1 00:06:03.088 15:10:30 rpc_client -- scripts/common.sh@365 -- # ver1[v]=1 00:06:03.088 15:10:30 rpc_client -- scripts/common.sh@366 -- # decimal 2 00:06:03.088 15:10:30 rpc_client -- scripts/common.sh@353 -- # local d=2 00:06:03.088 15:10:30 rpc_client -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:03.088 15:10:30 rpc_client -- scripts/common.sh@355 -- # echo 2 00:06:03.088 15:10:30 rpc_client -- scripts/common.sh@366 -- # ver2[v]=2 00:06:03.088 15:10:30 rpc_client -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:03.088 15:10:30 rpc_client -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:03.088 15:10:30 rpc_client -- scripts/common.sh@368 -- # return 0 00:06:03.088 15:10:30 rpc_client -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:03.088 15:10:30 rpc_client -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:06:03.088 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:03.088 --rc genhtml_branch_coverage=1 00:06:03.088 --rc genhtml_function_coverage=1 00:06:03.088 --rc genhtml_legend=1 00:06:03.088 --rc geninfo_all_blocks=1 00:06:03.088 --rc geninfo_unexecuted_blocks=1 00:06:03.088 00:06:03.088 ' 00:06:03.088 15:10:30 rpc_client -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:06:03.088 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:03.088 --rc genhtml_branch_coverage=1 00:06:03.088 --rc genhtml_function_coverage=1 00:06:03.088 --rc genhtml_legend=1 00:06:03.088 --rc geninfo_all_blocks=1 00:06:03.088 --rc geninfo_unexecuted_blocks=1 00:06:03.088 00:06:03.088 ' 00:06:03.088 15:10:30 rpc_client -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:06:03.088 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:03.088 --rc genhtml_branch_coverage=1 00:06:03.088 --rc genhtml_function_coverage=1 00:06:03.088 --rc genhtml_legend=1 00:06:03.088 --rc geninfo_all_blocks=1 00:06:03.088 --rc geninfo_unexecuted_blocks=1 00:06:03.088 00:06:03.088 ' 00:06:03.088 15:10:30 rpc_client -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:06:03.088 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:03.088 --rc genhtml_branch_coverage=1 00:06:03.088 --rc genhtml_function_coverage=1 00:06:03.088 --rc genhtml_legend=1 00:06:03.088 --rc geninfo_all_blocks=1 00:06:03.088 --rc geninfo_unexecuted_blocks=1 00:06:03.088 00:06:03.088 ' 00:06:03.088 15:10:30 rpc_client -- rpc_client/rpc_client.sh@10 -- # /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client_test 00:06:03.088 OK 00:06:03.088 15:10:31 rpc_client -- rpc_client/rpc_client.sh@12 -- # trap - SIGINT SIGTERM EXIT 00:06:03.088 00:06:03.088 real 0m0.300s 00:06:03.088 user 0m0.170s 00:06:03.088 sys 0m0.143s 00:06:03.088 15:10:31 rpc_client -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:03.088 15:10:31 rpc_client -- common/autotest_common.sh@10 -- # set +x 00:06:03.088 ************************************ 00:06:03.088 END TEST rpc_client 00:06:03.088 ************************************ 00:06:03.088 15:10:31 -- spdk/autotest.sh@159 -- # run_test json_config /home/vagrant/spdk_repo/spdk/test/json_config/json_config.sh 00:06:03.088 15:10:31 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:03.088 15:10:31 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:03.088 15:10:31 -- common/autotest_common.sh@10 -- # set +x 00:06:03.088 ************************************ 00:06:03.088 START TEST json_config 00:06:03.088 ************************************ 00:06:03.088 15:10:31 json_config -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/json_config/json_config.sh 00:06:03.354 15:10:31 json_config -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:06:03.354 15:10:31 json_config -- common/autotest_common.sh@1693 -- # lcov --version 00:06:03.354 15:10:31 json_config -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:06:03.354 15:10:31 json_config -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:06:03.354 15:10:31 json_config -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:03.354 15:10:31 json_config -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:03.354 15:10:31 json_config -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:03.354 15:10:31 json_config -- scripts/common.sh@336 -- # IFS=.-: 00:06:03.354 15:10:31 json_config -- scripts/common.sh@336 -- # read -ra ver1 00:06:03.354 15:10:31 json_config -- scripts/common.sh@337 -- # IFS=.-: 00:06:03.354 15:10:31 json_config -- scripts/common.sh@337 -- # read -ra ver2 00:06:03.354 15:10:31 json_config -- scripts/common.sh@338 -- # local 'op=<' 00:06:03.354 15:10:31 json_config -- scripts/common.sh@340 -- # ver1_l=2 00:06:03.354 15:10:31 json_config -- scripts/common.sh@341 -- # ver2_l=1 00:06:03.354 15:10:31 json_config -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:03.354 15:10:31 json_config -- scripts/common.sh@344 -- # case "$op" in 00:06:03.354 15:10:31 json_config -- scripts/common.sh@345 -- # : 1 00:06:03.354 15:10:31 json_config -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:03.354 15:10:31 json_config -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:03.354 15:10:31 json_config -- scripts/common.sh@365 -- # decimal 1 00:06:03.354 15:10:31 json_config -- scripts/common.sh@353 -- # local d=1 00:06:03.354 15:10:31 json_config -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:03.354 15:10:31 json_config -- scripts/common.sh@355 -- # echo 1 00:06:03.354 15:10:31 json_config -- scripts/common.sh@365 -- # ver1[v]=1 00:06:03.354 15:10:31 json_config -- scripts/common.sh@366 -- # decimal 2 00:06:03.354 15:10:31 json_config -- scripts/common.sh@353 -- # local d=2 00:06:03.354 15:10:31 json_config -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:03.354 15:10:31 json_config -- scripts/common.sh@355 -- # echo 2 00:06:03.354 15:10:31 json_config -- scripts/common.sh@366 -- # ver2[v]=2 00:06:03.354 15:10:31 json_config -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:03.354 15:10:31 json_config -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:03.354 15:10:31 json_config -- scripts/common.sh@368 -- # return 0 00:06:03.354 15:10:31 json_config -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:03.354 15:10:31 json_config -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:06:03.354 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:03.354 --rc genhtml_branch_coverage=1 00:06:03.354 --rc genhtml_function_coverage=1 00:06:03.354 --rc genhtml_legend=1 00:06:03.354 --rc geninfo_all_blocks=1 00:06:03.354 --rc geninfo_unexecuted_blocks=1 00:06:03.354 00:06:03.354 ' 00:06:03.354 15:10:31 json_config -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:06:03.354 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:03.354 --rc genhtml_branch_coverage=1 00:06:03.354 --rc genhtml_function_coverage=1 00:06:03.354 --rc genhtml_legend=1 00:06:03.354 --rc geninfo_all_blocks=1 00:06:03.354 --rc geninfo_unexecuted_blocks=1 00:06:03.354 00:06:03.354 ' 00:06:03.354 15:10:31 json_config -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:06:03.354 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:03.354 --rc genhtml_branch_coverage=1 00:06:03.354 --rc genhtml_function_coverage=1 00:06:03.354 --rc genhtml_legend=1 00:06:03.354 --rc geninfo_all_blocks=1 00:06:03.354 --rc geninfo_unexecuted_blocks=1 00:06:03.355 00:06:03.355 ' 00:06:03.355 15:10:31 json_config -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:06:03.355 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:03.355 --rc genhtml_branch_coverage=1 00:06:03.355 --rc genhtml_function_coverage=1 00:06:03.355 --rc genhtml_legend=1 00:06:03.355 --rc geninfo_all_blocks=1 00:06:03.355 --rc geninfo_unexecuted_blocks=1 00:06:03.355 00:06:03.355 ' 00:06:03.355 15:10:31 json_config -- json_config/json_config.sh@8 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:06:03.355 15:10:31 json_config -- nvmf/common.sh@7 -- # uname -s 00:06:03.355 15:10:31 json_config -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:06:03.355 15:10:31 json_config -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:06:03.355 15:10:31 json_config -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:06:03.355 15:10:31 json_config -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:06:03.355 15:10:31 json_config -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:06:03.355 15:10:31 json_config -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:06:03.355 15:10:31 json_config -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:06:03.355 15:10:31 json_config -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:06:03.355 15:10:31 json_config -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:06:03.355 15:10:31 json_config -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:06:03.355 15:10:31 json_config -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:d42b44b2-5d93-4f0c-9e37-1c47f6668401 00:06:03.355 15:10:31 json_config -- nvmf/common.sh@18 -- # NVME_HOSTID=d42b44b2-5d93-4f0c-9e37-1c47f6668401 00:06:03.355 15:10:31 json_config -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:06:03.355 15:10:31 json_config -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:06:03.355 15:10:31 json_config -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:06:03.355 15:10:31 json_config -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:06:03.355 15:10:31 json_config -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:06:03.355 15:10:31 json_config -- scripts/common.sh@15 -- # shopt -s extglob 00:06:03.355 15:10:31 json_config -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:06:03.355 15:10:31 json_config -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:06:03.355 15:10:31 json_config -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:06:03.355 15:10:31 json_config -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:03.355 15:10:31 json_config -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:03.355 15:10:31 json_config -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:03.355 15:10:31 json_config -- paths/export.sh@5 -- # export PATH 00:06:03.355 15:10:31 json_config -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:03.355 15:10:31 json_config -- nvmf/common.sh@51 -- # : 0 00:06:03.355 15:10:31 json_config -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:06:03.355 15:10:31 json_config -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:06:03.355 15:10:31 json_config -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:06:03.355 15:10:31 json_config -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:06:03.355 15:10:31 json_config -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:06:03.355 15:10:31 json_config -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:06:03.355 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:06:03.355 15:10:31 json_config -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:06:03.355 15:10:31 json_config -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:06:03.355 15:10:31 json_config -- nvmf/common.sh@55 -- # have_pci_nics=0 00:06:03.355 15:10:31 json_config -- json_config/json_config.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/json_config/common.sh 00:06:03.355 15:10:31 json_config -- json_config/json_config.sh@11 -- # [[ 0 -eq 1 ]] 00:06:03.355 15:10:31 json_config -- json_config/json_config.sh@15 -- # [[ 0 -ne 1 ]] 00:06:03.355 15:10:31 json_config -- json_config/json_config.sh@15 -- # [[ 0 -eq 1 ]] 00:06:03.355 15:10:31 json_config -- json_config/json_config.sh@26 -- # (( SPDK_TEST_BLOCKDEV + SPDK_TEST_ISCSI + SPDK_TEST_NVMF + SPDK_TEST_VHOST + SPDK_TEST_VHOST_INIT + SPDK_TEST_RBD == 0 )) 00:06:03.355 15:10:31 json_config -- json_config/json_config.sh@27 -- # echo 'WARNING: No tests are enabled so not running JSON configuration tests' 00:06:03.355 WARNING: No tests are enabled so not running JSON configuration tests 00:06:03.355 15:10:31 json_config -- json_config/json_config.sh@28 -- # exit 0 00:06:03.355 00:06:03.355 real 0m0.234s 00:06:03.355 user 0m0.146s 00:06:03.355 sys 0m0.092s 00:06:03.355 15:10:31 json_config -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:03.355 15:10:31 json_config -- common/autotest_common.sh@10 -- # set +x 00:06:03.355 ************************************ 00:06:03.355 END TEST json_config 00:06:03.355 ************************************ 00:06:03.355 15:10:31 -- spdk/autotest.sh@160 -- # run_test json_config_extra_key /home/vagrant/spdk_repo/spdk/test/json_config/json_config_extra_key.sh 00:06:03.355 15:10:31 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:03.355 15:10:31 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:03.355 15:10:31 -- common/autotest_common.sh@10 -- # set +x 00:06:03.355 ************************************ 00:06:03.355 START TEST json_config_extra_key 00:06:03.355 ************************************ 00:06:03.355 15:10:31 json_config_extra_key -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/json_config/json_config_extra_key.sh 00:06:03.624 15:10:31 json_config_extra_key -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:06:03.624 15:10:31 json_config_extra_key -- common/autotest_common.sh@1693 -- # lcov --version 00:06:03.624 15:10:31 json_config_extra_key -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:06:03.624 15:10:31 json_config_extra_key -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:06:03.624 15:10:31 json_config_extra_key -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:03.624 15:10:31 json_config_extra_key -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:03.624 15:10:31 json_config_extra_key -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:03.624 15:10:31 json_config_extra_key -- scripts/common.sh@336 -- # IFS=.-: 00:06:03.624 15:10:31 json_config_extra_key -- scripts/common.sh@336 -- # read -ra ver1 00:06:03.624 15:10:31 json_config_extra_key -- scripts/common.sh@337 -- # IFS=.-: 00:06:03.624 15:10:31 json_config_extra_key -- scripts/common.sh@337 -- # read -ra ver2 00:06:03.624 15:10:31 json_config_extra_key -- scripts/common.sh@338 -- # local 'op=<' 00:06:03.624 15:10:31 json_config_extra_key -- scripts/common.sh@340 -- # ver1_l=2 00:06:03.624 15:10:31 json_config_extra_key -- scripts/common.sh@341 -- # ver2_l=1 00:06:03.624 15:10:31 json_config_extra_key -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:03.624 15:10:31 json_config_extra_key -- scripts/common.sh@344 -- # case "$op" in 00:06:03.624 15:10:31 json_config_extra_key -- scripts/common.sh@345 -- # : 1 00:06:03.624 15:10:31 json_config_extra_key -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:03.624 15:10:31 json_config_extra_key -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:03.624 15:10:31 json_config_extra_key -- scripts/common.sh@365 -- # decimal 1 00:06:03.624 15:10:31 json_config_extra_key -- scripts/common.sh@353 -- # local d=1 00:06:03.624 15:10:31 json_config_extra_key -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:03.624 15:10:31 json_config_extra_key -- scripts/common.sh@355 -- # echo 1 00:06:03.624 15:10:31 json_config_extra_key -- scripts/common.sh@365 -- # ver1[v]=1 00:06:03.624 15:10:31 json_config_extra_key -- scripts/common.sh@366 -- # decimal 2 00:06:03.624 15:10:31 json_config_extra_key -- scripts/common.sh@353 -- # local d=2 00:06:03.624 15:10:31 json_config_extra_key -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:03.624 15:10:31 json_config_extra_key -- scripts/common.sh@355 -- # echo 2 00:06:03.625 15:10:31 json_config_extra_key -- scripts/common.sh@366 -- # ver2[v]=2 00:06:03.625 15:10:31 json_config_extra_key -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:03.625 15:10:31 json_config_extra_key -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:03.625 15:10:31 json_config_extra_key -- scripts/common.sh@368 -- # return 0 00:06:03.625 15:10:31 json_config_extra_key -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:03.625 15:10:31 json_config_extra_key -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:06:03.625 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:03.625 --rc genhtml_branch_coverage=1 00:06:03.625 --rc genhtml_function_coverage=1 00:06:03.625 --rc genhtml_legend=1 00:06:03.625 --rc geninfo_all_blocks=1 00:06:03.625 --rc geninfo_unexecuted_blocks=1 00:06:03.625 00:06:03.625 ' 00:06:03.625 15:10:31 json_config_extra_key -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:06:03.625 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:03.625 --rc genhtml_branch_coverage=1 00:06:03.625 --rc genhtml_function_coverage=1 00:06:03.625 --rc genhtml_legend=1 00:06:03.625 --rc geninfo_all_blocks=1 00:06:03.625 --rc geninfo_unexecuted_blocks=1 00:06:03.625 00:06:03.625 ' 00:06:03.625 15:10:31 json_config_extra_key -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:06:03.625 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:03.625 --rc genhtml_branch_coverage=1 00:06:03.625 --rc genhtml_function_coverage=1 00:06:03.625 --rc genhtml_legend=1 00:06:03.625 --rc geninfo_all_blocks=1 00:06:03.625 --rc geninfo_unexecuted_blocks=1 00:06:03.625 00:06:03.625 ' 00:06:03.625 15:10:31 json_config_extra_key -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:06:03.625 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:03.625 --rc genhtml_branch_coverage=1 00:06:03.625 --rc genhtml_function_coverage=1 00:06:03.625 --rc genhtml_legend=1 00:06:03.625 --rc geninfo_all_blocks=1 00:06:03.625 --rc geninfo_unexecuted_blocks=1 00:06:03.625 00:06:03.625 ' 00:06:03.625 15:10:31 json_config_extra_key -- json_config/json_config_extra_key.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:06:03.625 15:10:31 json_config_extra_key -- nvmf/common.sh@7 -- # uname -s 00:06:03.625 15:10:31 json_config_extra_key -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:06:03.625 15:10:31 json_config_extra_key -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:06:03.625 15:10:31 json_config_extra_key -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:06:03.625 15:10:31 json_config_extra_key -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:06:03.625 15:10:31 json_config_extra_key -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:06:03.625 15:10:31 json_config_extra_key -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:06:03.625 15:10:31 json_config_extra_key -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:06:03.625 15:10:31 json_config_extra_key -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:06:03.625 15:10:31 json_config_extra_key -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:06:03.625 15:10:31 json_config_extra_key -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:06:03.625 15:10:31 json_config_extra_key -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:d42b44b2-5d93-4f0c-9e37-1c47f6668401 00:06:03.625 15:10:31 json_config_extra_key -- nvmf/common.sh@18 -- # NVME_HOSTID=d42b44b2-5d93-4f0c-9e37-1c47f6668401 00:06:03.625 15:10:31 json_config_extra_key -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:06:03.625 15:10:31 json_config_extra_key -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:06:03.625 15:10:31 json_config_extra_key -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:06:03.625 15:10:31 json_config_extra_key -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:06:03.625 15:10:31 json_config_extra_key -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:06:03.625 15:10:31 json_config_extra_key -- scripts/common.sh@15 -- # shopt -s extglob 00:06:03.625 15:10:31 json_config_extra_key -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:06:03.625 15:10:31 json_config_extra_key -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:06:03.625 15:10:31 json_config_extra_key -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:06:03.625 15:10:31 json_config_extra_key -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:03.625 15:10:31 json_config_extra_key -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:03.625 15:10:31 json_config_extra_key -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:03.625 15:10:31 json_config_extra_key -- paths/export.sh@5 -- # export PATH 00:06:03.625 15:10:31 json_config_extra_key -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:03.625 15:10:31 json_config_extra_key -- nvmf/common.sh@51 -- # : 0 00:06:03.625 15:10:31 json_config_extra_key -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:06:03.625 15:10:31 json_config_extra_key -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:06:03.625 15:10:31 json_config_extra_key -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:06:03.625 15:10:31 json_config_extra_key -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:06:03.625 15:10:31 json_config_extra_key -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:06:03.625 15:10:31 json_config_extra_key -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:06:03.625 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:06:03.625 15:10:31 json_config_extra_key -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:06:03.625 15:10:31 json_config_extra_key -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:06:03.625 15:10:31 json_config_extra_key -- nvmf/common.sh@55 -- # have_pci_nics=0 00:06:03.625 15:10:31 json_config_extra_key -- json_config/json_config_extra_key.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/json_config/common.sh 00:06:03.625 15:10:31 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # app_pid=(['target']='') 00:06:03.625 15:10:31 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # declare -A app_pid 00:06:03.625 15:10:31 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # app_socket=(['target']='/var/tmp/spdk_tgt.sock') 00:06:03.625 15:10:31 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # declare -A app_socket 00:06:03.625 15:10:31 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # app_params=(['target']='-m 0x1 -s 1024') 00:06:03.625 15:10:31 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # declare -A app_params 00:06:03.625 15:10:31 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # configs_path=(['target']='/home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json') 00:06:03.625 15:10:31 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # declare -A configs_path 00:06:03.625 15:10:31 json_config_extra_key -- json_config/json_config_extra_key.sh@22 -- # trap 'on_error_exit "${FUNCNAME}" "${LINENO}"' ERR 00:06:03.625 15:10:31 json_config_extra_key -- json_config/json_config_extra_key.sh@24 -- # echo 'INFO: launching applications...' 00:06:03.625 INFO: launching applications... 00:06:03.625 15:10:31 json_config_extra_key -- json_config/json_config_extra_key.sh@25 -- # json_config_test_start_app target --json /home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json 00:06:03.625 15:10:31 json_config_extra_key -- json_config/common.sh@9 -- # local app=target 00:06:03.625 15:10:31 json_config_extra_key -- json_config/common.sh@10 -- # shift 00:06:03.625 15:10:31 json_config_extra_key -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:06:03.625 15:10:31 json_config_extra_key -- json_config/common.sh@13 -- # [[ -z '' ]] 00:06:03.625 15:10:31 json_config_extra_key -- json_config/common.sh@15 -- # local app_extra_params= 00:06:03.625 15:10:31 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:06:03.625 15:10:31 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:06:03.625 15:10:31 json_config_extra_key -- json_config/common.sh@22 -- # app_pid["$app"]=69933 00:06:03.625 15:10:31 json_config_extra_key -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:06:03.625 Waiting for target to run... 00:06:03.625 15:10:31 json_config_extra_key -- json_config/common.sh@25 -- # waitforlisten 69933 /var/tmp/spdk_tgt.sock 00:06:03.625 15:10:31 json_config_extra_key -- common/autotest_common.sh@835 -- # '[' -z 69933 ']' 00:06:03.625 15:10:31 json_config_extra_key -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:06:03.625 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:06:03.625 15:10:31 json_config_extra_key -- json_config/common.sh@21 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -s 1024 -r /var/tmp/spdk_tgt.sock --json /home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json 00:06:03.625 15:10:31 json_config_extra_key -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:03.625 15:10:31 json_config_extra_key -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:06:03.626 15:10:31 json_config_extra_key -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:03.626 15:10:31 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:06:03.886 [2024-11-27 15:10:31.757289] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:06:03.886 [2024-11-27 15:10:31.757865] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1024 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69933 ] 00:06:04.146 [2024-11-27 15:10:32.141786] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:04.146 [2024-11-27 15:10:32.161381] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:04.716 00:06:04.716 INFO: shutting down applications... 00:06:04.716 15:10:32 json_config_extra_key -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:04.716 15:10:32 json_config_extra_key -- common/autotest_common.sh@868 -- # return 0 00:06:04.716 15:10:32 json_config_extra_key -- json_config/common.sh@26 -- # echo '' 00:06:04.716 15:10:32 json_config_extra_key -- json_config/json_config_extra_key.sh@27 -- # echo 'INFO: shutting down applications...' 00:06:04.716 15:10:32 json_config_extra_key -- json_config/json_config_extra_key.sh@28 -- # json_config_test_shutdown_app target 00:06:04.716 15:10:32 json_config_extra_key -- json_config/common.sh@31 -- # local app=target 00:06:04.716 15:10:32 json_config_extra_key -- json_config/common.sh@34 -- # [[ -n 22 ]] 00:06:04.716 15:10:32 json_config_extra_key -- json_config/common.sh@35 -- # [[ -n 69933 ]] 00:06:04.716 15:10:32 json_config_extra_key -- json_config/common.sh@38 -- # kill -SIGINT 69933 00:06:04.716 15:10:32 json_config_extra_key -- json_config/common.sh@40 -- # (( i = 0 )) 00:06:04.716 15:10:32 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:06:04.716 15:10:32 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 69933 00:06:04.716 15:10:32 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:06:05.285 15:10:33 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:06:05.285 15:10:33 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:06:05.285 15:10:33 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 69933 00:06:05.285 15:10:33 json_config_extra_key -- json_config/common.sh@42 -- # app_pid["$app"]= 00:06:05.286 15:10:33 json_config_extra_key -- json_config/common.sh@43 -- # break 00:06:05.286 15:10:33 json_config_extra_key -- json_config/common.sh@48 -- # [[ -n '' ]] 00:06:05.286 15:10:33 json_config_extra_key -- json_config/common.sh@53 -- # echo 'SPDK target shutdown done' 00:06:05.286 SPDK target shutdown done 00:06:05.286 15:10:33 json_config_extra_key -- json_config/json_config_extra_key.sh@30 -- # echo Success 00:06:05.286 Success 00:06:05.286 00:06:05.286 real 0m1.683s 00:06:05.286 user 0m1.377s 00:06:05.286 sys 0m0.509s 00:06:05.286 15:10:33 json_config_extra_key -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:05.286 15:10:33 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:06:05.286 ************************************ 00:06:05.286 END TEST json_config_extra_key 00:06:05.286 ************************************ 00:06:05.286 15:10:33 -- spdk/autotest.sh@161 -- # run_test alias_rpc /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:06:05.286 15:10:33 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:05.286 15:10:33 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:05.286 15:10:33 -- common/autotest_common.sh@10 -- # set +x 00:06:05.286 ************************************ 00:06:05.286 START TEST alias_rpc 00:06:05.286 ************************************ 00:06:05.286 15:10:33 alias_rpc -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:06:05.286 * Looking for test storage... 00:06:05.286 * Found test storage at /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc 00:06:05.286 15:10:33 alias_rpc -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:06:05.286 15:10:33 alias_rpc -- common/autotest_common.sh@1693 -- # lcov --version 00:06:05.286 15:10:33 alias_rpc -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:06:05.286 15:10:33 alias_rpc -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:06:05.286 15:10:33 alias_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:05.286 15:10:33 alias_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:05.286 15:10:33 alias_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:05.286 15:10:33 alias_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:06:05.286 15:10:33 alias_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:06:05.286 15:10:33 alias_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:06:05.286 15:10:33 alias_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:06:05.286 15:10:33 alias_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:06:05.286 15:10:33 alias_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:06:05.286 15:10:33 alias_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:06:05.286 15:10:33 alias_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:05.545 15:10:33 alias_rpc -- scripts/common.sh@344 -- # case "$op" in 00:06:05.545 15:10:33 alias_rpc -- scripts/common.sh@345 -- # : 1 00:06:05.545 15:10:33 alias_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:05.545 15:10:33 alias_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:05.545 15:10:33 alias_rpc -- scripts/common.sh@365 -- # decimal 1 00:06:05.545 15:10:33 alias_rpc -- scripts/common.sh@353 -- # local d=1 00:06:05.545 15:10:33 alias_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:05.545 15:10:33 alias_rpc -- scripts/common.sh@355 -- # echo 1 00:06:05.545 15:10:33 alias_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:06:05.545 15:10:33 alias_rpc -- scripts/common.sh@366 -- # decimal 2 00:06:05.545 15:10:33 alias_rpc -- scripts/common.sh@353 -- # local d=2 00:06:05.545 15:10:33 alias_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:05.545 15:10:33 alias_rpc -- scripts/common.sh@355 -- # echo 2 00:06:05.545 15:10:33 alias_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:06:05.545 15:10:33 alias_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:05.545 15:10:33 alias_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:05.545 15:10:33 alias_rpc -- scripts/common.sh@368 -- # return 0 00:06:05.545 15:10:33 alias_rpc -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:05.545 15:10:33 alias_rpc -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:06:05.545 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:05.545 --rc genhtml_branch_coverage=1 00:06:05.545 --rc genhtml_function_coverage=1 00:06:05.545 --rc genhtml_legend=1 00:06:05.545 --rc geninfo_all_blocks=1 00:06:05.545 --rc geninfo_unexecuted_blocks=1 00:06:05.545 00:06:05.545 ' 00:06:05.545 15:10:33 alias_rpc -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:06:05.545 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:05.545 --rc genhtml_branch_coverage=1 00:06:05.545 --rc genhtml_function_coverage=1 00:06:05.545 --rc genhtml_legend=1 00:06:05.545 --rc geninfo_all_blocks=1 00:06:05.545 --rc geninfo_unexecuted_blocks=1 00:06:05.545 00:06:05.545 ' 00:06:05.545 15:10:33 alias_rpc -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:06:05.545 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:05.545 --rc genhtml_branch_coverage=1 00:06:05.545 --rc genhtml_function_coverage=1 00:06:05.545 --rc genhtml_legend=1 00:06:05.545 --rc geninfo_all_blocks=1 00:06:05.545 --rc geninfo_unexecuted_blocks=1 00:06:05.545 00:06:05.545 ' 00:06:05.545 15:10:33 alias_rpc -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:06:05.545 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:05.545 --rc genhtml_branch_coverage=1 00:06:05.545 --rc genhtml_function_coverage=1 00:06:05.545 --rc genhtml_legend=1 00:06:05.545 --rc geninfo_all_blocks=1 00:06:05.545 --rc geninfo_unexecuted_blocks=1 00:06:05.545 00:06:05.546 ' 00:06:05.546 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:05.546 15:10:33 alias_rpc -- alias_rpc/alias_rpc.sh@10 -- # trap 'killprocess $spdk_tgt_pid; exit 1' ERR 00:06:05.546 15:10:33 alias_rpc -- alias_rpc/alias_rpc.sh@13 -- # spdk_tgt_pid=70010 00:06:05.546 15:10:33 alias_rpc -- alias_rpc/alias_rpc.sh@12 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:05.546 15:10:33 alias_rpc -- alias_rpc/alias_rpc.sh@14 -- # waitforlisten 70010 00:06:05.546 15:10:33 alias_rpc -- common/autotest_common.sh@835 -- # '[' -z 70010 ']' 00:06:05.546 15:10:33 alias_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:05.546 15:10:33 alias_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:05.546 15:10:33 alias_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:05.546 15:10:33 alias_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:05.546 15:10:33 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:05.546 [2024-11-27 15:10:33.506171] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:06:05.546 [2024-11-27 15:10:33.506381] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70010 ] 00:06:05.805 [2024-11-27 15:10:33.678202] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:05.805 [2024-11-27 15:10:33.705304] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:06.376 15:10:34 alias_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:06.376 15:10:34 alias_rpc -- common/autotest_common.sh@868 -- # return 0 00:06:06.376 15:10:34 alias_rpc -- alias_rpc/alias_rpc.sh@17 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py load_config -i 00:06:06.636 15:10:34 alias_rpc -- alias_rpc/alias_rpc.sh@19 -- # killprocess 70010 00:06:06.636 15:10:34 alias_rpc -- common/autotest_common.sh@954 -- # '[' -z 70010 ']' 00:06:06.636 15:10:34 alias_rpc -- common/autotest_common.sh@958 -- # kill -0 70010 00:06:06.636 15:10:34 alias_rpc -- common/autotest_common.sh@959 -- # uname 00:06:06.636 15:10:34 alias_rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:06.636 15:10:34 alias_rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70010 00:06:06.636 15:10:34 alias_rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:06.636 15:10:34 alias_rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:06.636 15:10:34 alias_rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70010' 00:06:06.636 killing process with pid 70010 00:06:06.636 15:10:34 alias_rpc -- common/autotest_common.sh@973 -- # kill 70010 00:06:06.636 15:10:34 alias_rpc -- common/autotest_common.sh@978 -- # wait 70010 00:06:06.896 00:06:06.896 real 0m1.777s 00:06:06.896 user 0m1.814s 00:06:06.896 sys 0m0.496s 00:06:06.896 15:10:34 alias_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:06.896 15:10:34 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:06.896 ************************************ 00:06:06.896 END TEST alias_rpc 00:06:06.896 ************************************ 00:06:07.156 15:10:35 -- spdk/autotest.sh@163 -- # [[ 0 -eq 0 ]] 00:06:07.156 15:10:35 -- spdk/autotest.sh@164 -- # run_test spdkcli_tcp /home/vagrant/spdk_repo/spdk/test/spdkcli/tcp.sh 00:06:07.156 15:10:35 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:07.156 15:10:35 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:07.156 15:10:35 -- common/autotest_common.sh@10 -- # set +x 00:06:07.156 ************************************ 00:06:07.156 START TEST spdkcli_tcp 00:06:07.156 ************************************ 00:06:07.156 15:10:35 spdkcli_tcp -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/tcp.sh 00:06:07.156 * Looking for test storage... 00:06:07.156 * Found test storage at /home/vagrant/spdk_repo/spdk/test/spdkcli 00:06:07.156 15:10:35 spdkcli_tcp -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:06:07.156 15:10:35 spdkcli_tcp -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:06:07.156 15:10:35 spdkcli_tcp -- common/autotest_common.sh@1693 -- # lcov --version 00:06:07.156 15:10:35 spdkcli_tcp -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:06:07.156 15:10:35 spdkcli_tcp -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:07.156 15:10:35 spdkcli_tcp -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:07.156 15:10:35 spdkcli_tcp -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:07.156 15:10:35 spdkcli_tcp -- scripts/common.sh@336 -- # IFS=.-: 00:06:07.156 15:10:35 spdkcli_tcp -- scripts/common.sh@336 -- # read -ra ver1 00:06:07.156 15:10:35 spdkcli_tcp -- scripts/common.sh@337 -- # IFS=.-: 00:06:07.156 15:10:35 spdkcli_tcp -- scripts/common.sh@337 -- # read -ra ver2 00:06:07.156 15:10:35 spdkcli_tcp -- scripts/common.sh@338 -- # local 'op=<' 00:06:07.156 15:10:35 spdkcli_tcp -- scripts/common.sh@340 -- # ver1_l=2 00:06:07.156 15:10:35 spdkcli_tcp -- scripts/common.sh@341 -- # ver2_l=1 00:06:07.156 15:10:35 spdkcli_tcp -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:07.156 15:10:35 spdkcli_tcp -- scripts/common.sh@344 -- # case "$op" in 00:06:07.156 15:10:35 spdkcli_tcp -- scripts/common.sh@345 -- # : 1 00:06:07.156 15:10:35 spdkcli_tcp -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:07.156 15:10:35 spdkcli_tcp -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:07.156 15:10:35 spdkcli_tcp -- scripts/common.sh@365 -- # decimal 1 00:06:07.156 15:10:35 spdkcli_tcp -- scripts/common.sh@353 -- # local d=1 00:06:07.156 15:10:35 spdkcli_tcp -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:07.156 15:10:35 spdkcli_tcp -- scripts/common.sh@355 -- # echo 1 00:06:07.156 15:10:35 spdkcli_tcp -- scripts/common.sh@365 -- # ver1[v]=1 00:06:07.156 15:10:35 spdkcli_tcp -- scripts/common.sh@366 -- # decimal 2 00:06:07.156 15:10:35 spdkcli_tcp -- scripts/common.sh@353 -- # local d=2 00:06:07.156 15:10:35 spdkcli_tcp -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:07.156 15:10:35 spdkcli_tcp -- scripts/common.sh@355 -- # echo 2 00:06:07.156 15:10:35 spdkcli_tcp -- scripts/common.sh@366 -- # ver2[v]=2 00:06:07.156 15:10:35 spdkcli_tcp -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:07.156 15:10:35 spdkcli_tcp -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:07.156 15:10:35 spdkcli_tcp -- scripts/common.sh@368 -- # return 0 00:06:07.156 15:10:35 spdkcli_tcp -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:07.156 15:10:35 spdkcli_tcp -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:06:07.156 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:07.156 --rc genhtml_branch_coverage=1 00:06:07.156 --rc genhtml_function_coverage=1 00:06:07.156 --rc genhtml_legend=1 00:06:07.156 --rc geninfo_all_blocks=1 00:06:07.156 --rc geninfo_unexecuted_blocks=1 00:06:07.156 00:06:07.156 ' 00:06:07.156 15:10:35 spdkcli_tcp -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:06:07.156 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:07.156 --rc genhtml_branch_coverage=1 00:06:07.156 --rc genhtml_function_coverage=1 00:06:07.156 --rc genhtml_legend=1 00:06:07.156 --rc geninfo_all_blocks=1 00:06:07.156 --rc geninfo_unexecuted_blocks=1 00:06:07.156 00:06:07.156 ' 00:06:07.156 15:10:35 spdkcli_tcp -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:06:07.156 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:07.156 --rc genhtml_branch_coverage=1 00:06:07.156 --rc genhtml_function_coverage=1 00:06:07.156 --rc genhtml_legend=1 00:06:07.156 --rc geninfo_all_blocks=1 00:06:07.156 --rc geninfo_unexecuted_blocks=1 00:06:07.156 00:06:07.156 ' 00:06:07.156 15:10:35 spdkcli_tcp -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:06:07.156 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:07.156 --rc genhtml_branch_coverage=1 00:06:07.156 --rc genhtml_function_coverage=1 00:06:07.156 --rc genhtml_legend=1 00:06:07.156 --rc geninfo_all_blocks=1 00:06:07.156 --rc geninfo_unexecuted_blocks=1 00:06:07.156 00:06:07.156 ' 00:06:07.156 15:10:35 spdkcli_tcp -- spdkcli/tcp.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:06:07.156 15:10:35 spdkcli_tcp -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:06:07.157 15:10:35 spdkcli_tcp -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:06:07.157 15:10:35 spdkcli_tcp -- spdkcli/tcp.sh@18 -- # IP_ADDRESS=127.0.0.1 00:06:07.157 15:10:35 spdkcli_tcp -- spdkcli/tcp.sh@19 -- # PORT=9998 00:06:07.157 15:10:35 spdkcli_tcp -- spdkcli/tcp.sh@21 -- # trap 'err_cleanup; exit 1' SIGINT SIGTERM EXIT 00:06:07.157 15:10:35 spdkcli_tcp -- spdkcli/tcp.sh@23 -- # timing_enter run_spdk_tgt_tcp 00:06:07.157 15:10:35 spdkcli_tcp -- common/autotest_common.sh@726 -- # xtrace_disable 00:06:07.157 15:10:35 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:06:07.416 15:10:35 spdkcli_tcp -- spdkcli/tcp.sh@25 -- # spdk_tgt_pid=70095 00:06:07.416 15:10:35 spdkcli_tcp -- spdkcli/tcp.sh@24 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x3 -p 0 00:06:07.416 15:10:35 spdkcli_tcp -- spdkcli/tcp.sh@27 -- # waitforlisten 70095 00:06:07.416 15:10:35 spdkcli_tcp -- common/autotest_common.sh@835 -- # '[' -z 70095 ']' 00:06:07.416 15:10:35 spdkcli_tcp -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:07.416 15:10:35 spdkcli_tcp -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:07.416 15:10:35 spdkcli_tcp -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:07.416 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:07.416 15:10:35 spdkcli_tcp -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:07.416 15:10:35 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:06:07.416 [2024-11-27 15:10:35.361762] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:06:07.416 [2024-11-27 15:10:35.362018] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70095 ] 00:06:07.675 [2024-11-27 15:10:35.531178] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:07.675 [2024-11-27 15:10:35.561338] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:07.675 [2024-11-27 15:10:35.561429] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:08.245 15:10:36 spdkcli_tcp -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:08.245 15:10:36 spdkcli_tcp -- common/autotest_common.sh@868 -- # return 0 00:06:08.245 15:10:36 spdkcli_tcp -- spdkcli/tcp.sh@31 -- # socat_pid=70112 00:06:08.245 15:10:36 spdkcli_tcp -- spdkcli/tcp.sh@30 -- # socat TCP-LISTEN:9998 UNIX-CONNECT:/var/tmp/spdk.sock 00:06:08.245 15:10:36 spdkcli_tcp -- spdkcli/tcp.sh@33 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -r 100 -t 2 -s 127.0.0.1 -p 9998 rpc_get_methods 00:06:08.505 [ 00:06:08.505 "bdev_malloc_delete", 00:06:08.505 "bdev_malloc_create", 00:06:08.506 "bdev_null_resize", 00:06:08.506 "bdev_null_delete", 00:06:08.506 "bdev_null_create", 00:06:08.506 "bdev_nvme_cuse_unregister", 00:06:08.506 "bdev_nvme_cuse_register", 00:06:08.506 "bdev_opal_new_user", 00:06:08.506 "bdev_opal_set_lock_state", 00:06:08.506 "bdev_opal_delete", 00:06:08.506 "bdev_opal_get_info", 00:06:08.506 "bdev_opal_create", 00:06:08.506 "bdev_nvme_opal_revert", 00:06:08.506 "bdev_nvme_opal_init", 00:06:08.506 "bdev_nvme_send_cmd", 00:06:08.506 "bdev_nvme_set_keys", 00:06:08.506 "bdev_nvme_get_path_iostat", 00:06:08.506 "bdev_nvme_get_mdns_discovery_info", 00:06:08.506 "bdev_nvme_stop_mdns_discovery", 00:06:08.506 "bdev_nvme_start_mdns_discovery", 00:06:08.506 "bdev_nvme_set_multipath_policy", 00:06:08.506 "bdev_nvme_set_preferred_path", 00:06:08.506 "bdev_nvme_get_io_paths", 00:06:08.506 "bdev_nvme_remove_error_injection", 00:06:08.506 "bdev_nvme_add_error_injection", 00:06:08.506 "bdev_nvme_get_discovery_info", 00:06:08.506 "bdev_nvme_stop_discovery", 00:06:08.506 "bdev_nvme_start_discovery", 00:06:08.506 "bdev_nvme_get_controller_health_info", 00:06:08.506 "bdev_nvme_disable_controller", 00:06:08.506 "bdev_nvme_enable_controller", 00:06:08.506 "bdev_nvme_reset_controller", 00:06:08.506 "bdev_nvme_get_transport_statistics", 00:06:08.506 "bdev_nvme_apply_firmware", 00:06:08.506 "bdev_nvme_detach_controller", 00:06:08.506 "bdev_nvme_get_controllers", 00:06:08.506 "bdev_nvme_attach_controller", 00:06:08.506 "bdev_nvme_set_hotplug", 00:06:08.506 "bdev_nvme_set_options", 00:06:08.506 "bdev_passthru_delete", 00:06:08.506 "bdev_passthru_create", 00:06:08.506 "bdev_lvol_set_parent_bdev", 00:06:08.506 "bdev_lvol_set_parent", 00:06:08.506 "bdev_lvol_check_shallow_copy", 00:06:08.506 "bdev_lvol_start_shallow_copy", 00:06:08.506 "bdev_lvol_grow_lvstore", 00:06:08.506 "bdev_lvol_get_lvols", 00:06:08.506 "bdev_lvol_get_lvstores", 00:06:08.506 "bdev_lvol_delete", 00:06:08.506 "bdev_lvol_set_read_only", 00:06:08.506 "bdev_lvol_resize", 00:06:08.506 "bdev_lvol_decouple_parent", 00:06:08.506 "bdev_lvol_inflate", 00:06:08.506 "bdev_lvol_rename", 00:06:08.506 "bdev_lvol_clone_bdev", 00:06:08.506 "bdev_lvol_clone", 00:06:08.506 "bdev_lvol_snapshot", 00:06:08.506 "bdev_lvol_create", 00:06:08.506 "bdev_lvol_delete_lvstore", 00:06:08.506 "bdev_lvol_rename_lvstore", 00:06:08.506 "bdev_lvol_create_lvstore", 00:06:08.506 "bdev_raid_set_options", 00:06:08.506 "bdev_raid_remove_base_bdev", 00:06:08.506 "bdev_raid_add_base_bdev", 00:06:08.506 "bdev_raid_delete", 00:06:08.506 "bdev_raid_create", 00:06:08.506 "bdev_raid_get_bdevs", 00:06:08.506 "bdev_error_inject_error", 00:06:08.506 "bdev_error_delete", 00:06:08.506 "bdev_error_create", 00:06:08.506 "bdev_split_delete", 00:06:08.506 "bdev_split_create", 00:06:08.506 "bdev_delay_delete", 00:06:08.506 "bdev_delay_create", 00:06:08.506 "bdev_delay_update_latency", 00:06:08.506 "bdev_zone_block_delete", 00:06:08.506 "bdev_zone_block_create", 00:06:08.506 "blobfs_create", 00:06:08.506 "blobfs_detect", 00:06:08.506 "blobfs_set_cache_size", 00:06:08.506 "bdev_aio_delete", 00:06:08.506 "bdev_aio_rescan", 00:06:08.506 "bdev_aio_create", 00:06:08.506 "bdev_ftl_set_property", 00:06:08.506 "bdev_ftl_get_properties", 00:06:08.506 "bdev_ftl_get_stats", 00:06:08.506 "bdev_ftl_unmap", 00:06:08.506 "bdev_ftl_unload", 00:06:08.506 "bdev_ftl_delete", 00:06:08.506 "bdev_ftl_load", 00:06:08.506 "bdev_ftl_create", 00:06:08.506 "bdev_virtio_attach_controller", 00:06:08.506 "bdev_virtio_scsi_get_devices", 00:06:08.506 "bdev_virtio_detach_controller", 00:06:08.506 "bdev_virtio_blk_set_hotplug", 00:06:08.506 "bdev_iscsi_delete", 00:06:08.506 "bdev_iscsi_create", 00:06:08.506 "bdev_iscsi_set_options", 00:06:08.506 "accel_error_inject_error", 00:06:08.506 "ioat_scan_accel_module", 00:06:08.506 "dsa_scan_accel_module", 00:06:08.506 "iaa_scan_accel_module", 00:06:08.506 "keyring_file_remove_key", 00:06:08.506 "keyring_file_add_key", 00:06:08.506 "keyring_linux_set_options", 00:06:08.506 "fsdev_aio_delete", 00:06:08.506 "fsdev_aio_create", 00:06:08.506 "iscsi_get_histogram", 00:06:08.506 "iscsi_enable_histogram", 00:06:08.506 "iscsi_set_options", 00:06:08.506 "iscsi_get_auth_groups", 00:06:08.506 "iscsi_auth_group_remove_secret", 00:06:08.506 "iscsi_auth_group_add_secret", 00:06:08.506 "iscsi_delete_auth_group", 00:06:08.506 "iscsi_create_auth_group", 00:06:08.506 "iscsi_set_discovery_auth", 00:06:08.506 "iscsi_get_options", 00:06:08.506 "iscsi_target_node_request_logout", 00:06:08.506 "iscsi_target_node_set_redirect", 00:06:08.506 "iscsi_target_node_set_auth", 00:06:08.506 "iscsi_target_node_add_lun", 00:06:08.506 "iscsi_get_stats", 00:06:08.506 "iscsi_get_connections", 00:06:08.506 "iscsi_portal_group_set_auth", 00:06:08.506 "iscsi_start_portal_group", 00:06:08.506 "iscsi_delete_portal_group", 00:06:08.506 "iscsi_create_portal_group", 00:06:08.506 "iscsi_get_portal_groups", 00:06:08.506 "iscsi_delete_target_node", 00:06:08.506 "iscsi_target_node_remove_pg_ig_maps", 00:06:08.506 "iscsi_target_node_add_pg_ig_maps", 00:06:08.506 "iscsi_create_target_node", 00:06:08.506 "iscsi_get_target_nodes", 00:06:08.506 "iscsi_delete_initiator_group", 00:06:08.506 "iscsi_initiator_group_remove_initiators", 00:06:08.506 "iscsi_initiator_group_add_initiators", 00:06:08.506 "iscsi_create_initiator_group", 00:06:08.506 "iscsi_get_initiator_groups", 00:06:08.506 "nvmf_set_crdt", 00:06:08.506 "nvmf_set_config", 00:06:08.506 "nvmf_set_max_subsystems", 00:06:08.506 "nvmf_stop_mdns_prr", 00:06:08.506 "nvmf_publish_mdns_prr", 00:06:08.506 "nvmf_subsystem_get_listeners", 00:06:08.506 "nvmf_subsystem_get_qpairs", 00:06:08.506 "nvmf_subsystem_get_controllers", 00:06:08.506 "nvmf_get_stats", 00:06:08.506 "nvmf_get_transports", 00:06:08.506 "nvmf_create_transport", 00:06:08.506 "nvmf_get_targets", 00:06:08.506 "nvmf_delete_target", 00:06:08.506 "nvmf_create_target", 00:06:08.506 "nvmf_subsystem_allow_any_host", 00:06:08.506 "nvmf_subsystem_set_keys", 00:06:08.506 "nvmf_subsystem_remove_host", 00:06:08.506 "nvmf_subsystem_add_host", 00:06:08.506 "nvmf_ns_remove_host", 00:06:08.506 "nvmf_ns_add_host", 00:06:08.506 "nvmf_subsystem_remove_ns", 00:06:08.506 "nvmf_subsystem_set_ns_ana_group", 00:06:08.506 "nvmf_subsystem_add_ns", 00:06:08.506 "nvmf_subsystem_listener_set_ana_state", 00:06:08.506 "nvmf_discovery_get_referrals", 00:06:08.506 "nvmf_discovery_remove_referral", 00:06:08.506 "nvmf_discovery_add_referral", 00:06:08.506 "nvmf_subsystem_remove_listener", 00:06:08.506 "nvmf_subsystem_add_listener", 00:06:08.506 "nvmf_delete_subsystem", 00:06:08.506 "nvmf_create_subsystem", 00:06:08.506 "nvmf_get_subsystems", 00:06:08.506 "env_dpdk_get_mem_stats", 00:06:08.506 "nbd_get_disks", 00:06:08.506 "nbd_stop_disk", 00:06:08.506 "nbd_start_disk", 00:06:08.506 "ublk_recover_disk", 00:06:08.506 "ublk_get_disks", 00:06:08.506 "ublk_stop_disk", 00:06:08.506 "ublk_start_disk", 00:06:08.506 "ublk_destroy_target", 00:06:08.506 "ublk_create_target", 00:06:08.506 "virtio_blk_create_transport", 00:06:08.506 "virtio_blk_get_transports", 00:06:08.506 "vhost_controller_set_coalescing", 00:06:08.506 "vhost_get_controllers", 00:06:08.506 "vhost_delete_controller", 00:06:08.506 "vhost_create_blk_controller", 00:06:08.506 "vhost_scsi_controller_remove_target", 00:06:08.506 "vhost_scsi_controller_add_target", 00:06:08.506 "vhost_start_scsi_controller", 00:06:08.506 "vhost_create_scsi_controller", 00:06:08.506 "thread_set_cpumask", 00:06:08.506 "scheduler_set_options", 00:06:08.506 "framework_get_governor", 00:06:08.506 "framework_get_scheduler", 00:06:08.506 "framework_set_scheduler", 00:06:08.506 "framework_get_reactors", 00:06:08.506 "thread_get_io_channels", 00:06:08.506 "thread_get_pollers", 00:06:08.506 "thread_get_stats", 00:06:08.506 "framework_monitor_context_switch", 00:06:08.506 "spdk_kill_instance", 00:06:08.506 "log_enable_timestamps", 00:06:08.506 "log_get_flags", 00:06:08.506 "log_clear_flag", 00:06:08.506 "log_set_flag", 00:06:08.506 "log_get_level", 00:06:08.506 "log_set_level", 00:06:08.506 "log_get_print_level", 00:06:08.506 "log_set_print_level", 00:06:08.506 "framework_enable_cpumask_locks", 00:06:08.506 "framework_disable_cpumask_locks", 00:06:08.506 "framework_wait_init", 00:06:08.506 "framework_start_init", 00:06:08.506 "scsi_get_devices", 00:06:08.506 "bdev_get_histogram", 00:06:08.506 "bdev_enable_histogram", 00:06:08.506 "bdev_set_qos_limit", 00:06:08.506 "bdev_set_qd_sampling_period", 00:06:08.506 "bdev_get_bdevs", 00:06:08.506 "bdev_reset_iostat", 00:06:08.506 "bdev_get_iostat", 00:06:08.506 "bdev_examine", 00:06:08.506 "bdev_wait_for_examine", 00:06:08.506 "bdev_set_options", 00:06:08.506 "accel_get_stats", 00:06:08.506 "accel_set_options", 00:06:08.506 "accel_set_driver", 00:06:08.506 "accel_crypto_key_destroy", 00:06:08.506 "accel_crypto_keys_get", 00:06:08.506 "accel_crypto_key_create", 00:06:08.506 "accel_assign_opc", 00:06:08.506 "accel_get_module_info", 00:06:08.506 "accel_get_opc_assignments", 00:06:08.506 "vmd_rescan", 00:06:08.506 "vmd_remove_device", 00:06:08.506 "vmd_enable", 00:06:08.506 "sock_get_default_impl", 00:06:08.506 "sock_set_default_impl", 00:06:08.506 "sock_impl_set_options", 00:06:08.506 "sock_impl_get_options", 00:06:08.506 "iobuf_get_stats", 00:06:08.506 "iobuf_set_options", 00:06:08.506 "keyring_get_keys", 00:06:08.506 "framework_get_pci_devices", 00:06:08.506 "framework_get_config", 00:06:08.506 "framework_get_subsystems", 00:06:08.506 "fsdev_set_opts", 00:06:08.506 "fsdev_get_opts", 00:06:08.506 "trace_get_info", 00:06:08.506 "trace_get_tpoint_group_mask", 00:06:08.506 "trace_disable_tpoint_group", 00:06:08.506 "trace_enable_tpoint_group", 00:06:08.506 "trace_clear_tpoint_mask", 00:06:08.506 "trace_set_tpoint_mask", 00:06:08.506 "notify_get_notifications", 00:06:08.506 "notify_get_types", 00:06:08.506 "spdk_get_version", 00:06:08.506 "rpc_get_methods" 00:06:08.506 ] 00:06:08.506 15:10:36 spdkcli_tcp -- spdkcli/tcp.sh@35 -- # timing_exit run_spdk_tgt_tcp 00:06:08.506 15:10:36 spdkcli_tcp -- common/autotest_common.sh@732 -- # xtrace_disable 00:06:08.506 15:10:36 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:06:08.506 15:10:36 spdkcli_tcp -- spdkcli/tcp.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:06:08.506 15:10:36 spdkcli_tcp -- spdkcli/tcp.sh@38 -- # killprocess 70095 00:06:08.506 15:10:36 spdkcli_tcp -- common/autotest_common.sh@954 -- # '[' -z 70095 ']' 00:06:08.506 15:10:36 spdkcli_tcp -- common/autotest_common.sh@958 -- # kill -0 70095 00:06:08.506 15:10:36 spdkcli_tcp -- common/autotest_common.sh@959 -- # uname 00:06:08.506 15:10:36 spdkcli_tcp -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:08.506 15:10:36 spdkcli_tcp -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70095 00:06:08.506 15:10:36 spdkcli_tcp -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:08.506 15:10:36 spdkcli_tcp -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:08.506 15:10:36 spdkcli_tcp -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70095' 00:06:08.506 killing process with pid 70095 00:06:08.506 15:10:36 spdkcli_tcp -- common/autotest_common.sh@973 -- # kill 70095 00:06:08.506 15:10:36 spdkcli_tcp -- common/autotest_common.sh@978 -- # wait 70095 00:06:08.766 00:06:08.766 real 0m1.812s 00:06:08.766 user 0m2.991s 00:06:08.766 sys 0m0.574s 00:06:08.766 ************************************ 00:06:08.766 END TEST spdkcli_tcp 00:06:08.766 ************************************ 00:06:08.766 15:10:36 spdkcli_tcp -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:08.766 15:10:36 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:06:09.025 15:10:36 -- spdk/autotest.sh@167 -- # run_test dpdk_mem_utility /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:06:09.025 15:10:36 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:09.025 15:10:36 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:09.025 15:10:36 -- common/autotest_common.sh@10 -- # set +x 00:06:09.025 ************************************ 00:06:09.025 START TEST dpdk_mem_utility 00:06:09.025 ************************************ 00:06:09.025 15:10:36 dpdk_mem_utility -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:06:09.025 * Looking for test storage... 00:06:09.025 * Found test storage at /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility 00:06:09.025 15:10:37 dpdk_mem_utility -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:06:09.025 15:10:37 dpdk_mem_utility -- common/autotest_common.sh@1693 -- # lcov --version 00:06:09.025 15:10:37 dpdk_mem_utility -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:06:09.025 15:10:37 dpdk_mem_utility -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:06:09.025 15:10:37 dpdk_mem_utility -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:09.026 15:10:37 dpdk_mem_utility -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:09.026 15:10:37 dpdk_mem_utility -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:09.026 15:10:37 dpdk_mem_utility -- scripts/common.sh@336 -- # IFS=.-: 00:06:09.026 15:10:37 dpdk_mem_utility -- scripts/common.sh@336 -- # read -ra ver1 00:06:09.026 15:10:37 dpdk_mem_utility -- scripts/common.sh@337 -- # IFS=.-: 00:06:09.026 15:10:37 dpdk_mem_utility -- scripts/common.sh@337 -- # read -ra ver2 00:06:09.026 15:10:37 dpdk_mem_utility -- scripts/common.sh@338 -- # local 'op=<' 00:06:09.026 15:10:37 dpdk_mem_utility -- scripts/common.sh@340 -- # ver1_l=2 00:06:09.026 15:10:37 dpdk_mem_utility -- scripts/common.sh@341 -- # ver2_l=1 00:06:09.026 15:10:37 dpdk_mem_utility -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:09.026 15:10:37 dpdk_mem_utility -- scripts/common.sh@344 -- # case "$op" in 00:06:09.026 15:10:37 dpdk_mem_utility -- scripts/common.sh@345 -- # : 1 00:06:09.026 15:10:37 dpdk_mem_utility -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:09.026 15:10:37 dpdk_mem_utility -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:09.026 15:10:37 dpdk_mem_utility -- scripts/common.sh@365 -- # decimal 1 00:06:09.026 15:10:37 dpdk_mem_utility -- scripts/common.sh@353 -- # local d=1 00:06:09.026 15:10:37 dpdk_mem_utility -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:09.026 15:10:37 dpdk_mem_utility -- scripts/common.sh@355 -- # echo 1 00:06:09.026 15:10:37 dpdk_mem_utility -- scripts/common.sh@365 -- # ver1[v]=1 00:06:09.026 15:10:37 dpdk_mem_utility -- scripts/common.sh@366 -- # decimal 2 00:06:09.026 15:10:37 dpdk_mem_utility -- scripts/common.sh@353 -- # local d=2 00:06:09.026 15:10:37 dpdk_mem_utility -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:09.026 15:10:37 dpdk_mem_utility -- scripts/common.sh@355 -- # echo 2 00:06:09.286 15:10:37 dpdk_mem_utility -- scripts/common.sh@366 -- # ver2[v]=2 00:06:09.286 15:10:37 dpdk_mem_utility -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:09.286 15:10:37 dpdk_mem_utility -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:09.286 15:10:37 dpdk_mem_utility -- scripts/common.sh@368 -- # return 0 00:06:09.286 15:10:37 dpdk_mem_utility -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:09.286 15:10:37 dpdk_mem_utility -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:06:09.286 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:09.286 --rc genhtml_branch_coverage=1 00:06:09.286 --rc genhtml_function_coverage=1 00:06:09.286 --rc genhtml_legend=1 00:06:09.286 --rc geninfo_all_blocks=1 00:06:09.286 --rc geninfo_unexecuted_blocks=1 00:06:09.286 00:06:09.286 ' 00:06:09.286 15:10:37 dpdk_mem_utility -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:06:09.286 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:09.286 --rc genhtml_branch_coverage=1 00:06:09.286 --rc genhtml_function_coverage=1 00:06:09.286 --rc genhtml_legend=1 00:06:09.286 --rc geninfo_all_blocks=1 00:06:09.286 --rc geninfo_unexecuted_blocks=1 00:06:09.286 00:06:09.286 ' 00:06:09.286 15:10:37 dpdk_mem_utility -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:06:09.286 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:09.286 --rc genhtml_branch_coverage=1 00:06:09.286 --rc genhtml_function_coverage=1 00:06:09.286 --rc genhtml_legend=1 00:06:09.286 --rc geninfo_all_blocks=1 00:06:09.286 --rc geninfo_unexecuted_blocks=1 00:06:09.286 00:06:09.286 ' 00:06:09.286 15:10:37 dpdk_mem_utility -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:06:09.286 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:09.286 --rc genhtml_branch_coverage=1 00:06:09.286 --rc genhtml_function_coverage=1 00:06:09.286 --rc genhtml_legend=1 00:06:09.286 --rc geninfo_all_blocks=1 00:06:09.286 --rc geninfo_unexecuted_blocks=1 00:06:09.286 00:06:09.286 ' 00:06:09.286 15:10:37 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@10 -- # MEM_SCRIPT=/home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py 00:06:09.286 15:10:37 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@13 -- # spdkpid=70195 00:06:09.286 15:10:37 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@12 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:09.286 15:10:37 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@15 -- # waitforlisten 70195 00:06:09.286 15:10:37 dpdk_mem_utility -- common/autotest_common.sh@835 -- # '[' -z 70195 ']' 00:06:09.286 15:10:37 dpdk_mem_utility -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:09.286 15:10:37 dpdk_mem_utility -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:09.286 15:10:37 dpdk_mem_utility -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:09.286 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:09.286 15:10:37 dpdk_mem_utility -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:09.286 15:10:37 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:06:09.286 [2024-11-27 15:10:37.232335] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:06:09.286 [2024-11-27 15:10:37.232550] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70195 ] 00:06:09.546 [2024-11-27 15:10:37.402636] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:09.546 [2024-11-27 15:10:37.430699] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:10.120 15:10:38 dpdk_mem_utility -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:10.120 15:10:38 dpdk_mem_utility -- common/autotest_common.sh@868 -- # return 0 00:06:10.120 15:10:38 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@17 -- # trap 'killprocess $spdkpid' SIGINT SIGTERM EXIT 00:06:10.120 15:10:38 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@19 -- # rpc_cmd env_dpdk_get_mem_stats 00:06:10.120 15:10:38 dpdk_mem_utility -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:10.120 15:10:38 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:06:10.120 { 00:06:10.120 "filename": "/tmp/spdk_mem_dump.txt" 00:06:10.120 } 00:06:10.120 15:10:38 dpdk_mem_utility -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:10.120 15:10:38 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@21 -- # /home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py 00:06:10.120 DPDK memory size 818.000000 MiB in 1 heap(s) 00:06:10.120 1 heaps totaling size 818.000000 MiB 00:06:10.120 size: 818.000000 MiB heap id: 0 00:06:10.120 end heaps---------- 00:06:10.120 9 mempools totaling size 603.782043 MiB 00:06:10.120 size: 212.674988 MiB name: PDU_immediate_data_Pool 00:06:10.120 size: 158.602051 MiB name: PDU_data_out_Pool 00:06:10.120 size: 100.555481 MiB name: bdev_io_70195 00:06:10.121 size: 50.003479 MiB name: msgpool_70195 00:06:10.121 size: 36.509338 MiB name: fsdev_io_70195 00:06:10.121 size: 21.763794 MiB name: PDU_Pool 00:06:10.121 size: 19.513306 MiB name: SCSI_TASK_Pool 00:06:10.121 size: 4.133484 MiB name: evtpool_70195 00:06:10.121 size: 0.026123 MiB name: Session_Pool 00:06:10.121 end mempools------- 00:06:10.121 6 memzones totaling size 4.142822 MiB 00:06:10.121 size: 1.000366 MiB name: RG_ring_0_70195 00:06:10.121 size: 1.000366 MiB name: RG_ring_1_70195 00:06:10.121 size: 1.000366 MiB name: RG_ring_4_70195 00:06:10.121 size: 1.000366 MiB name: RG_ring_5_70195 00:06:10.121 size: 0.125366 MiB name: RG_ring_2_70195 00:06:10.121 size: 0.015991 MiB name: RG_ring_3_70195 00:06:10.121 end memzones------- 00:06:10.121 15:10:38 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@23 -- # /home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py -m 0 00:06:10.121 heap id: 0 total size: 818.000000 MiB number of busy elements: 313 number of free elements: 15 00:06:10.121 list of free elements. size: 10.803223 MiB 00:06:10.121 element at address: 0x200019200000 with size: 0.999878 MiB 00:06:10.121 element at address: 0x200019400000 with size: 0.999878 MiB 00:06:10.121 element at address: 0x200032000000 with size: 0.994446 MiB 00:06:10.121 element at address: 0x200000400000 with size: 0.993958 MiB 00:06:10.121 element at address: 0x200006400000 with size: 0.959839 MiB 00:06:10.121 element at address: 0x200012c00000 with size: 0.944275 MiB 00:06:10.121 element at address: 0x200019600000 with size: 0.936584 MiB 00:06:10.121 element at address: 0x200000200000 with size: 0.717346 MiB 00:06:10.121 element at address: 0x20001ae00000 with size: 0.567871 MiB 00:06:10.121 element at address: 0x20000a600000 with size: 0.488892 MiB 00:06:10.121 element at address: 0x200000c00000 with size: 0.486267 MiB 00:06:10.121 element at address: 0x200019800000 with size: 0.485657 MiB 00:06:10.121 element at address: 0x200003e00000 with size: 0.480286 MiB 00:06:10.121 element at address: 0x200028200000 with size: 0.396301 MiB 00:06:10.121 element at address: 0x200000800000 with size: 0.351746 MiB 00:06:10.121 list of standard malloc elements. size: 199.267883 MiB 00:06:10.121 element at address: 0x20000a7fff80 with size: 132.000122 MiB 00:06:10.121 element at address: 0x2000065fff80 with size: 64.000122 MiB 00:06:10.121 element at address: 0x2000192fff80 with size: 1.000122 MiB 00:06:10.121 element at address: 0x2000194fff80 with size: 1.000122 MiB 00:06:10.121 element at address: 0x2000196fff80 with size: 1.000122 MiB 00:06:10.121 element at address: 0x2000003d9f00 with size: 0.140747 MiB 00:06:10.121 element at address: 0x2000196eff00 with size: 0.062622 MiB 00:06:10.121 element at address: 0x2000003fdf80 with size: 0.007935 MiB 00:06:10.121 element at address: 0x2000196efdc0 with size: 0.000305 MiB 00:06:10.121 element at address: 0x2000002d7c40 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000003d9e40 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000004fe740 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000004fe800 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000004fe8c0 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000004fe980 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000004fea40 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000004feb00 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000004febc0 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000004fec80 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000004fed40 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000004fee00 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000004feec0 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000004fef80 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000004ff040 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000004ff100 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000004ff1c0 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000004ff280 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000004ff340 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000004ff400 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000004ff4c0 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000004ff580 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000004ff640 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000004ff700 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000004ff7c0 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000004ff880 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000004ff940 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000004ffa00 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000004ffac0 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000004ffcc0 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000004ffd80 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000004ffe40 with size: 0.000183 MiB 00:06:10.121 element at address: 0x20000085a0c0 with size: 0.000183 MiB 00:06:10.121 element at address: 0x20000085a2c0 with size: 0.000183 MiB 00:06:10.121 element at address: 0x20000085e580 with size: 0.000183 MiB 00:06:10.121 element at address: 0x20000087e840 with size: 0.000183 MiB 00:06:10.121 element at address: 0x20000087e900 with size: 0.000183 MiB 00:06:10.121 element at address: 0x20000087e9c0 with size: 0.000183 MiB 00:06:10.121 element at address: 0x20000087ea80 with size: 0.000183 MiB 00:06:10.121 element at address: 0x20000087eb40 with size: 0.000183 MiB 00:06:10.121 element at address: 0x20000087ec00 with size: 0.000183 MiB 00:06:10.121 element at address: 0x20000087ecc0 with size: 0.000183 MiB 00:06:10.121 element at address: 0x20000087ed80 with size: 0.000183 MiB 00:06:10.121 element at address: 0x20000087ee40 with size: 0.000183 MiB 00:06:10.121 element at address: 0x20000087ef00 with size: 0.000183 MiB 00:06:10.121 element at address: 0x20000087efc0 with size: 0.000183 MiB 00:06:10.121 element at address: 0x20000087f080 with size: 0.000183 MiB 00:06:10.121 element at address: 0x20000087f140 with size: 0.000183 MiB 00:06:10.121 element at address: 0x20000087f200 with size: 0.000183 MiB 00:06:10.121 element at address: 0x20000087f2c0 with size: 0.000183 MiB 00:06:10.121 element at address: 0x20000087f380 with size: 0.000183 MiB 00:06:10.121 element at address: 0x20000087f440 with size: 0.000183 MiB 00:06:10.121 element at address: 0x20000087f500 with size: 0.000183 MiB 00:06:10.121 element at address: 0x20000087f5c0 with size: 0.000183 MiB 00:06:10.121 element at address: 0x20000087f680 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000008ff940 with size: 0.000183 MiB 00:06:10.121 element at address: 0x2000008ffb40 with size: 0.000183 MiB 00:06:10.121 element at address: 0x200000c7c7c0 with size: 0.000183 MiB 00:06:10.121 element at address: 0x200000c7c880 with size: 0.000183 MiB 00:06:10.121 element at address: 0x200000c7c940 with size: 0.000183 MiB 00:06:10.121 element at address: 0x200000c7ca00 with size: 0.000183 MiB 00:06:10.121 element at address: 0x200000c7cac0 with size: 0.000183 MiB 00:06:10.121 element at address: 0x200000c7cb80 with size: 0.000183 MiB 00:06:10.121 element at address: 0x200000c7cc40 with size: 0.000183 MiB 00:06:10.121 element at address: 0x200000c7cd00 with size: 0.000183 MiB 00:06:10.121 element at address: 0x200000c7cdc0 with size: 0.000183 MiB 00:06:10.121 element at address: 0x200000c7ce80 with size: 0.000183 MiB 00:06:10.121 element at address: 0x200000c7cf40 with size: 0.000183 MiB 00:06:10.121 element at address: 0x200000c7d000 with size: 0.000183 MiB 00:06:10.121 element at address: 0x200000c7d0c0 with size: 0.000183 MiB 00:06:10.121 element at address: 0x200000c7d180 with size: 0.000183 MiB 00:06:10.121 element at address: 0x200000c7d240 with size: 0.000183 MiB 00:06:10.121 element at address: 0x200000c7d300 with size: 0.000183 MiB 00:06:10.121 element at address: 0x200000c7d3c0 with size: 0.000183 MiB 00:06:10.121 element at address: 0x200000c7d480 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7d540 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7d600 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7d6c0 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7d780 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7d840 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7d900 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7d9c0 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7da80 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7db40 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7dc00 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7dcc0 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7dd80 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7de40 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7df00 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7dfc0 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7e080 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7e140 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7e200 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7e2c0 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7e380 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7e440 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7e500 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7e5c0 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7e680 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7e740 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7e800 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7e8c0 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7e980 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7ea40 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7eb00 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7ebc0 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7ec80 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000c7ed40 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000cff000 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200000cff0c0 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200003e7af40 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200003e7b000 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200003e7b0c0 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200003e7b180 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200003e7b240 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200003e7b300 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200003e7b3c0 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200003e7b480 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200003e7b540 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200003e7b600 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200003e7b6c0 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200003efb980 with size: 0.000183 MiB 00:06:10.122 element at address: 0x2000064fdd80 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20000a67d280 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20000a67d340 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20000a67d400 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20000a67d4c0 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20000a67d580 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20000a67d640 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20000a67d700 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20000a67d7c0 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20000a67d880 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20000a67d940 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20000a67da00 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20000a67dac0 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20000a6fdd80 with size: 0.000183 MiB 00:06:10.122 element at address: 0x200012cf1bc0 with size: 0.000183 MiB 00:06:10.122 element at address: 0x2000196efc40 with size: 0.000183 MiB 00:06:10.122 element at address: 0x2000196efd00 with size: 0.000183 MiB 00:06:10.122 element at address: 0x2000198bc740 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae91600 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae916c0 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae91780 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae91840 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae91900 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae919c0 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae91a80 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae91b40 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae91c00 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae91cc0 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae91d80 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae91e40 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae91f00 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae91fc0 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae92080 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae92140 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae92200 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae922c0 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae92380 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae92440 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae92500 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae925c0 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae92680 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae92740 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae92800 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae928c0 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae92980 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae92a40 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae92b00 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae92bc0 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae92c80 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae92d40 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae92e00 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae92ec0 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae92f80 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae93040 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae93100 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae931c0 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae93280 with size: 0.000183 MiB 00:06:10.122 element at address: 0x20001ae93340 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae93400 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae934c0 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae93580 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae93640 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae93700 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae937c0 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae93880 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae93940 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae93a00 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae93ac0 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae93b80 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae93c40 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae93d00 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae93dc0 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae93e80 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae93f40 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae94000 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae940c0 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae94180 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae94240 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae94300 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae943c0 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae94480 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae94540 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae94600 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae946c0 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae94780 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae94840 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae94900 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae949c0 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae94a80 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae94b40 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae94c00 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae94cc0 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae94d80 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae94e40 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae94f00 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae94fc0 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae95080 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae95140 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae95200 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae952c0 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae95380 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20001ae95440 with size: 0.000183 MiB 00:06:10.123 element at address: 0x200028265740 with size: 0.000183 MiB 00:06:10.123 element at address: 0x200028265800 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826c400 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826c600 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826c6c0 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826c780 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826c840 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826c900 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826c9c0 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826ca80 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826cb40 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826cc00 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826ccc0 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826cd80 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826ce40 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826cf00 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826cfc0 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826d080 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826d140 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826d200 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826d2c0 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826d380 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826d440 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826d500 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826d5c0 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826d680 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826d740 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826d800 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826d8c0 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826d980 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826da40 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826db00 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826dbc0 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826dc80 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826dd40 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826de00 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826dec0 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826df80 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826e040 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826e100 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826e1c0 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826e280 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826e340 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826e400 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826e4c0 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826e580 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826e640 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826e700 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826e7c0 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826e880 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826e940 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826ea00 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826eac0 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826eb80 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826ec40 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826ed00 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826edc0 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826ee80 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826ef40 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826f000 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826f0c0 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826f180 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826f240 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826f300 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826f3c0 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826f480 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826f540 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826f600 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826f6c0 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826f780 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826f840 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826f900 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826f9c0 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826fa80 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826fb40 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826fc00 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826fcc0 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826fd80 with size: 0.000183 MiB 00:06:10.123 element at address: 0x20002826fe40 with size: 0.000183 MiB 00:06:10.124 element at address: 0x20002826ff00 with size: 0.000183 MiB 00:06:10.124 list of memzone associated elements. size: 607.928894 MiB 00:06:10.124 element at address: 0x20001ae95500 with size: 211.416748 MiB 00:06:10.124 associated memzone info: size: 211.416626 MiB name: MP_PDU_immediate_data_Pool_0 00:06:10.124 element at address: 0x20002826ffc0 with size: 157.562561 MiB 00:06:10.124 associated memzone info: size: 157.562439 MiB name: MP_PDU_data_out_Pool_0 00:06:10.124 element at address: 0x200012df1e80 with size: 100.055054 MiB 00:06:10.124 associated memzone info: size: 100.054932 MiB name: MP_bdev_io_70195_0 00:06:10.124 element at address: 0x200000dff380 with size: 48.003052 MiB 00:06:10.124 associated memzone info: size: 48.002930 MiB name: MP_msgpool_70195_0 00:06:10.124 element at address: 0x200003ffdb80 with size: 36.008911 MiB 00:06:10.124 associated memzone info: size: 36.008789 MiB name: MP_fsdev_io_70195_0 00:06:10.124 element at address: 0x2000199be940 with size: 20.255554 MiB 00:06:10.124 associated memzone info: size: 20.255432 MiB name: MP_PDU_Pool_0 00:06:10.124 element at address: 0x2000321feb40 with size: 18.005066 MiB 00:06:10.124 associated memzone info: size: 18.004944 MiB name: MP_SCSI_TASK_Pool_0 00:06:10.124 element at address: 0x2000004fff00 with size: 3.000244 MiB 00:06:10.124 associated memzone info: size: 3.000122 MiB name: MP_evtpool_70195_0 00:06:10.124 element at address: 0x2000009ffe00 with size: 2.000488 MiB 00:06:10.124 associated memzone info: size: 2.000366 MiB name: RG_MP_msgpool_70195 00:06:10.124 element at address: 0x2000002d7d00 with size: 1.008118 MiB 00:06:10.124 associated memzone info: size: 1.007996 MiB name: MP_evtpool_70195 00:06:10.124 element at address: 0x20000a6fde40 with size: 1.008118 MiB 00:06:10.124 associated memzone info: size: 1.007996 MiB name: MP_PDU_Pool 00:06:10.124 element at address: 0x2000198bc800 with size: 1.008118 MiB 00:06:10.124 associated memzone info: size: 1.007996 MiB name: MP_PDU_immediate_data_Pool 00:06:10.124 element at address: 0x2000064fde40 with size: 1.008118 MiB 00:06:10.124 associated memzone info: size: 1.007996 MiB name: MP_PDU_data_out_Pool 00:06:10.124 element at address: 0x200003efba40 with size: 1.008118 MiB 00:06:10.124 associated memzone info: size: 1.007996 MiB name: MP_SCSI_TASK_Pool 00:06:10.124 element at address: 0x200000cff180 with size: 1.000488 MiB 00:06:10.124 associated memzone info: size: 1.000366 MiB name: RG_ring_0_70195 00:06:10.124 element at address: 0x2000008ffc00 with size: 1.000488 MiB 00:06:10.124 associated memzone info: size: 1.000366 MiB name: RG_ring_1_70195 00:06:10.124 element at address: 0x200012cf1c80 with size: 1.000488 MiB 00:06:10.124 associated memzone info: size: 1.000366 MiB name: RG_ring_4_70195 00:06:10.124 element at address: 0x2000320fe940 with size: 1.000488 MiB 00:06:10.124 associated memzone info: size: 1.000366 MiB name: RG_ring_5_70195 00:06:10.124 element at address: 0x20000087f740 with size: 0.500488 MiB 00:06:10.124 associated memzone info: size: 0.500366 MiB name: RG_MP_fsdev_io_70195 00:06:10.124 element at address: 0x200000c7ee00 with size: 0.500488 MiB 00:06:10.124 associated memzone info: size: 0.500366 MiB name: RG_MP_bdev_io_70195 00:06:10.124 element at address: 0x20000a67db80 with size: 0.500488 MiB 00:06:10.124 associated memzone info: size: 0.500366 MiB name: RG_MP_PDU_Pool 00:06:10.124 element at address: 0x200003e7b780 with size: 0.500488 MiB 00:06:10.124 associated memzone info: size: 0.500366 MiB name: RG_MP_SCSI_TASK_Pool 00:06:10.124 element at address: 0x20001987c540 with size: 0.250488 MiB 00:06:10.124 associated memzone info: size: 0.250366 MiB name: RG_MP_PDU_immediate_data_Pool 00:06:10.124 element at address: 0x2000002b7a40 with size: 0.125488 MiB 00:06:10.124 associated memzone info: size: 0.125366 MiB name: RG_MP_evtpool_70195 00:06:10.124 element at address: 0x20000085e640 with size: 0.125488 MiB 00:06:10.124 associated memzone info: size: 0.125366 MiB name: RG_ring_2_70195 00:06:10.124 element at address: 0x2000064f5b80 with size: 0.031738 MiB 00:06:10.124 associated memzone info: size: 0.031616 MiB name: RG_MP_PDU_data_out_Pool 00:06:10.124 element at address: 0x2000282658c0 with size: 0.023743 MiB 00:06:10.124 associated memzone info: size: 0.023621 MiB name: MP_Session_Pool_0 00:06:10.124 element at address: 0x20000085a380 with size: 0.016113 MiB 00:06:10.124 associated memzone info: size: 0.015991 MiB name: RG_ring_3_70195 00:06:10.124 element at address: 0x20002826ba00 with size: 0.002441 MiB 00:06:10.124 associated memzone info: size: 0.002319 MiB name: RG_MP_Session_Pool 00:06:10.124 element at address: 0x2000004ffb80 with size: 0.000305 MiB 00:06:10.124 associated memzone info: size: 0.000183 MiB name: MP_msgpool_70195 00:06:10.124 element at address: 0x2000008ffa00 with size: 0.000305 MiB 00:06:10.124 associated memzone info: size: 0.000183 MiB name: MP_fsdev_io_70195 00:06:10.124 element at address: 0x20000085a180 with size: 0.000305 MiB 00:06:10.124 associated memzone info: size: 0.000183 MiB name: MP_bdev_io_70195 00:06:10.124 element at address: 0x20002826c4c0 with size: 0.000305 MiB 00:06:10.124 associated memzone info: size: 0.000183 MiB name: MP_Session_Pool 00:06:10.124 15:10:38 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@25 -- # trap - SIGINT SIGTERM EXIT 00:06:10.124 15:10:38 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@26 -- # killprocess 70195 00:06:10.124 15:10:38 dpdk_mem_utility -- common/autotest_common.sh@954 -- # '[' -z 70195 ']' 00:06:10.124 15:10:38 dpdk_mem_utility -- common/autotest_common.sh@958 -- # kill -0 70195 00:06:10.385 15:10:38 dpdk_mem_utility -- common/autotest_common.sh@959 -- # uname 00:06:10.385 15:10:38 dpdk_mem_utility -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:10.385 15:10:38 dpdk_mem_utility -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70195 00:06:10.385 15:10:38 dpdk_mem_utility -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:10.385 15:10:38 dpdk_mem_utility -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:10.385 15:10:38 dpdk_mem_utility -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70195' 00:06:10.385 killing process with pid 70195 00:06:10.385 15:10:38 dpdk_mem_utility -- common/autotest_common.sh@973 -- # kill 70195 00:06:10.385 15:10:38 dpdk_mem_utility -- common/autotest_common.sh@978 -- # wait 70195 00:06:10.645 00:06:10.645 real 0m1.721s 00:06:10.645 user 0m1.714s 00:06:10.645 sys 0m0.512s 00:06:10.645 ************************************ 00:06:10.645 END TEST dpdk_mem_utility 00:06:10.645 ************************************ 00:06:10.645 15:10:38 dpdk_mem_utility -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:10.645 15:10:38 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:06:10.645 15:10:38 -- spdk/autotest.sh@168 -- # run_test event /home/vagrant/spdk_repo/spdk/test/event/event.sh 00:06:10.645 15:10:38 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:10.645 15:10:38 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:10.645 15:10:38 -- common/autotest_common.sh@10 -- # set +x 00:06:10.645 ************************************ 00:06:10.645 START TEST event 00:06:10.645 ************************************ 00:06:10.645 15:10:38 event -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/event.sh 00:06:10.905 * Looking for test storage... 00:06:10.905 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event 00:06:10.905 15:10:38 event -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:06:10.905 15:10:38 event -- common/autotest_common.sh@1693 -- # lcov --version 00:06:10.905 15:10:38 event -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:06:10.905 15:10:38 event -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:06:10.905 15:10:38 event -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:10.905 15:10:38 event -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:10.905 15:10:38 event -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:10.905 15:10:38 event -- scripts/common.sh@336 -- # IFS=.-: 00:06:10.905 15:10:38 event -- scripts/common.sh@336 -- # read -ra ver1 00:06:10.905 15:10:38 event -- scripts/common.sh@337 -- # IFS=.-: 00:06:10.905 15:10:38 event -- scripts/common.sh@337 -- # read -ra ver2 00:06:10.905 15:10:38 event -- scripts/common.sh@338 -- # local 'op=<' 00:06:10.905 15:10:38 event -- scripts/common.sh@340 -- # ver1_l=2 00:06:10.905 15:10:38 event -- scripts/common.sh@341 -- # ver2_l=1 00:06:10.905 15:10:38 event -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:10.905 15:10:38 event -- scripts/common.sh@344 -- # case "$op" in 00:06:10.905 15:10:38 event -- scripts/common.sh@345 -- # : 1 00:06:10.905 15:10:38 event -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:10.905 15:10:38 event -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:10.905 15:10:38 event -- scripts/common.sh@365 -- # decimal 1 00:06:10.905 15:10:38 event -- scripts/common.sh@353 -- # local d=1 00:06:10.905 15:10:38 event -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:10.905 15:10:38 event -- scripts/common.sh@355 -- # echo 1 00:06:10.905 15:10:38 event -- scripts/common.sh@365 -- # ver1[v]=1 00:06:10.905 15:10:38 event -- scripts/common.sh@366 -- # decimal 2 00:06:10.905 15:10:38 event -- scripts/common.sh@353 -- # local d=2 00:06:10.905 15:10:38 event -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:10.905 15:10:38 event -- scripts/common.sh@355 -- # echo 2 00:06:10.905 15:10:38 event -- scripts/common.sh@366 -- # ver2[v]=2 00:06:10.905 15:10:38 event -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:10.905 15:10:38 event -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:10.905 15:10:38 event -- scripts/common.sh@368 -- # return 0 00:06:10.905 15:10:38 event -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:10.905 15:10:38 event -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:06:10.906 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:10.906 --rc genhtml_branch_coverage=1 00:06:10.906 --rc genhtml_function_coverage=1 00:06:10.906 --rc genhtml_legend=1 00:06:10.906 --rc geninfo_all_blocks=1 00:06:10.906 --rc geninfo_unexecuted_blocks=1 00:06:10.906 00:06:10.906 ' 00:06:10.906 15:10:38 event -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:06:10.906 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:10.906 --rc genhtml_branch_coverage=1 00:06:10.906 --rc genhtml_function_coverage=1 00:06:10.906 --rc genhtml_legend=1 00:06:10.906 --rc geninfo_all_blocks=1 00:06:10.906 --rc geninfo_unexecuted_blocks=1 00:06:10.906 00:06:10.906 ' 00:06:10.906 15:10:38 event -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:06:10.906 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:10.906 --rc genhtml_branch_coverage=1 00:06:10.906 --rc genhtml_function_coverage=1 00:06:10.906 --rc genhtml_legend=1 00:06:10.906 --rc geninfo_all_blocks=1 00:06:10.906 --rc geninfo_unexecuted_blocks=1 00:06:10.906 00:06:10.906 ' 00:06:10.906 15:10:38 event -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:06:10.906 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:10.906 --rc genhtml_branch_coverage=1 00:06:10.906 --rc genhtml_function_coverage=1 00:06:10.906 --rc genhtml_legend=1 00:06:10.906 --rc geninfo_all_blocks=1 00:06:10.906 --rc geninfo_unexecuted_blocks=1 00:06:10.906 00:06:10.906 ' 00:06:10.906 15:10:38 event -- event/event.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:06:10.906 15:10:38 event -- bdev/nbd_common.sh@6 -- # set -e 00:06:10.906 15:10:38 event -- event/event.sh@45 -- # run_test event_perf /home/vagrant/spdk_repo/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:06:10.906 15:10:38 event -- common/autotest_common.sh@1105 -- # '[' 6 -le 1 ']' 00:06:10.906 15:10:38 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:10.906 15:10:38 event -- common/autotest_common.sh@10 -- # set +x 00:06:10.906 ************************************ 00:06:10.906 START TEST event_perf 00:06:10.906 ************************************ 00:06:10.906 15:10:38 event.event_perf -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:06:10.906 Running I/O for 1 seconds...[2024-11-27 15:10:38.999989] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:06:10.906 [2024-11-27 15:10:39.000139] [ DPDK EAL parameters: event_perf --no-shconf -c 0xF --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70270 ] 00:06:11.165 [2024-11-27 15:10:39.177111] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:06:11.165 [2024-11-27 15:10:39.210614] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:11.165 [2024-11-27 15:10:39.210824] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:06:11.165 [2024-11-27 15:10:39.210878] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:11.165 [2024-11-27 15:10:39.211038] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:06:12.542 Running I/O for 1 seconds... 00:06:12.542 lcore 0: 105826 00:06:12.542 lcore 1: 105825 00:06:12.542 lcore 2: 105828 00:06:12.542 lcore 3: 105826 00:06:12.542 done. 00:06:12.542 00:06:12.542 real 0m1.323s 00:06:12.542 user 0m4.071s 00:06:12.542 sys 0m0.127s 00:06:12.542 15:10:40 event.event_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:12.542 ************************************ 00:06:12.542 END TEST event_perf 00:06:12.542 ************************************ 00:06:12.542 15:10:40 event.event_perf -- common/autotest_common.sh@10 -- # set +x 00:06:12.542 15:10:40 event -- event/event.sh@46 -- # run_test event_reactor /home/vagrant/spdk_repo/spdk/test/event/reactor/reactor -t 1 00:06:12.542 15:10:40 event -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:06:12.542 15:10:40 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:12.542 15:10:40 event -- common/autotest_common.sh@10 -- # set +x 00:06:12.542 ************************************ 00:06:12.542 START TEST event_reactor 00:06:12.542 ************************************ 00:06:12.542 15:10:40 event.event_reactor -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/reactor/reactor -t 1 00:06:12.542 [2024-11-27 15:10:40.389449] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:06:12.542 [2024-11-27 15:10:40.389586] [ DPDK EAL parameters: reactor --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70315 ] 00:06:12.542 [2024-11-27 15:10:40.542998] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:12.542 [2024-11-27 15:10:40.571080] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:13.934 test_start 00:06:13.934 oneshot 00:06:13.934 tick 100 00:06:13.934 tick 100 00:06:13.934 tick 250 00:06:13.934 tick 100 00:06:13.934 tick 100 00:06:13.934 tick 100 00:06:13.934 tick 250 00:06:13.934 tick 500 00:06:13.934 tick 100 00:06:13.934 tick 100 00:06:13.934 tick 250 00:06:13.934 tick 100 00:06:13.934 tick 100 00:06:13.934 test_end 00:06:13.934 00:06:13.934 real 0m1.290s 00:06:13.934 user 0m1.096s 00:06:13.934 sys 0m0.088s 00:06:13.934 ************************************ 00:06:13.934 END TEST event_reactor 00:06:13.934 ************************************ 00:06:13.934 15:10:41 event.event_reactor -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:13.934 15:10:41 event.event_reactor -- common/autotest_common.sh@10 -- # set +x 00:06:13.934 15:10:41 event -- event/event.sh@47 -- # run_test event_reactor_perf /home/vagrant/spdk_repo/spdk/test/event/reactor_perf/reactor_perf -t 1 00:06:13.934 15:10:41 event -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:06:13.934 15:10:41 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:13.934 15:10:41 event -- common/autotest_common.sh@10 -- # set +x 00:06:13.934 ************************************ 00:06:13.934 START TEST event_reactor_perf 00:06:13.934 ************************************ 00:06:13.934 15:10:41 event.event_reactor_perf -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/reactor_perf/reactor_perf -t 1 00:06:13.934 [2024-11-27 15:10:41.749393] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:06:13.934 [2024-11-27 15:10:41.749548] [ DPDK EAL parameters: reactor_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70346 ] 00:06:13.934 [2024-11-27 15:10:41.919274] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:13.934 [2024-11-27 15:10:41.946175] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:15.316 test_start 00:06:15.316 test_end 00:06:15.316 Performance: 412647 events per second 00:06:15.316 00:06:15.316 real 0m1.305s 00:06:15.316 user 0m1.118s 00:06:15.316 sys 0m0.080s 00:06:15.316 15:10:43 event.event_reactor_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:15.316 15:10:43 event.event_reactor_perf -- common/autotest_common.sh@10 -- # set +x 00:06:15.316 ************************************ 00:06:15.316 END TEST event_reactor_perf 00:06:15.316 ************************************ 00:06:15.316 15:10:43 event -- event/event.sh@49 -- # uname -s 00:06:15.316 15:10:43 event -- event/event.sh@49 -- # '[' Linux = Linux ']' 00:06:15.316 15:10:43 event -- event/event.sh@50 -- # run_test event_scheduler /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler.sh 00:06:15.316 15:10:43 event -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:15.316 15:10:43 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:15.316 15:10:43 event -- common/autotest_common.sh@10 -- # set +x 00:06:15.316 ************************************ 00:06:15.316 START TEST event_scheduler 00:06:15.316 ************************************ 00:06:15.316 15:10:43 event.event_scheduler -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler.sh 00:06:15.316 * Looking for test storage... 00:06:15.316 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event/scheduler 00:06:15.316 15:10:43 event.event_scheduler -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:06:15.316 15:10:43 event.event_scheduler -- common/autotest_common.sh@1693 -- # lcov --version 00:06:15.316 15:10:43 event.event_scheduler -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:06:15.316 15:10:43 event.event_scheduler -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:06:15.316 15:10:43 event.event_scheduler -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:15.316 15:10:43 event.event_scheduler -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:15.316 15:10:43 event.event_scheduler -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:15.316 15:10:43 event.event_scheduler -- scripts/common.sh@336 -- # IFS=.-: 00:06:15.316 15:10:43 event.event_scheduler -- scripts/common.sh@336 -- # read -ra ver1 00:06:15.316 15:10:43 event.event_scheduler -- scripts/common.sh@337 -- # IFS=.-: 00:06:15.316 15:10:43 event.event_scheduler -- scripts/common.sh@337 -- # read -ra ver2 00:06:15.316 15:10:43 event.event_scheduler -- scripts/common.sh@338 -- # local 'op=<' 00:06:15.316 15:10:43 event.event_scheduler -- scripts/common.sh@340 -- # ver1_l=2 00:06:15.316 15:10:43 event.event_scheduler -- scripts/common.sh@341 -- # ver2_l=1 00:06:15.316 15:10:43 event.event_scheduler -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:15.316 15:10:43 event.event_scheduler -- scripts/common.sh@344 -- # case "$op" in 00:06:15.316 15:10:43 event.event_scheduler -- scripts/common.sh@345 -- # : 1 00:06:15.316 15:10:43 event.event_scheduler -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:15.316 15:10:43 event.event_scheduler -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:15.316 15:10:43 event.event_scheduler -- scripts/common.sh@365 -- # decimal 1 00:06:15.316 15:10:43 event.event_scheduler -- scripts/common.sh@353 -- # local d=1 00:06:15.316 15:10:43 event.event_scheduler -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:15.316 15:10:43 event.event_scheduler -- scripts/common.sh@355 -- # echo 1 00:06:15.316 15:10:43 event.event_scheduler -- scripts/common.sh@365 -- # ver1[v]=1 00:06:15.316 15:10:43 event.event_scheduler -- scripts/common.sh@366 -- # decimal 2 00:06:15.316 15:10:43 event.event_scheduler -- scripts/common.sh@353 -- # local d=2 00:06:15.316 15:10:43 event.event_scheduler -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:15.316 15:10:43 event.event_scheduler -- scripts/common.sh@355 -- # echo 2 00:06:15.316 15:10:43 event.event_scheduler -- scripts/common.sh@366 -- # ver2[v]=2 00:06:15.316 15:10:43 event.event_scheduler -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:15.316 15:10:43 event.event_scheduler -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:15.316 15:10:43 event.event_scheduler -- scripts/common.sh@368 -- # return 0 00:06:15.316 15:10:43 event.event_scheduler -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:15.316 15:10:43 event.event_scheduler -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:06:15.316 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:15.316 --rc genhtml_branch_coverage=1 00:06:15.316 --rc genhtml_function_coverage=1 00:06:15.316 --rc genhtml_legend=1 00:06:15.316 --rc geninfo_all_blocks=1 00:06:15.316 --rc geninfo_unexecuted_blocks=1 00:06:15.316 00:06:15.316 ' 00:06:15.316 15:10:43 event.event_scheduler -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:06:15.316 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:15.316 --rc genhtml_branch_coverage=1 00:06:15.316 --rc genhtml_function_coverage=1 00:06:15.316 --rc genhtml_legend=1 00:06:15.316 --rc geninfo_all_blocks=1 00:06:15.316 --rc geninfo_unexecuted_blocks=1 00:06:15.316 00:06:15.316 ' 00:06:15.316 15:10:43 event.event_scheduler -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:06:15.317 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:15.317 --rc genhtml_branch_coverage=1 00:06:15.317 --rc genhtml_function_coverage=1 00:06:15.317 --rc genhtml_legend=1 00:06:15.317 --rc geninfo_all_blocks=1 00:06:15.317 --rc geninfo_unexecuted_blocks=1 00:06:15.317 00:06:15.317 ' 00:06:15.317 15:10:43 event.event_scheduler -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:06:15.317 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:15.317 --rc genhtml_branch_coverage=1 00:06:15.317 --rc genhtml_function_coverage=1 00:06:15.317 --rc genhtml_legend=1 00:06:15.317 --rc geninfo_all_blocks=1 00:06:15.317 --rc geninfo_unexecuted_blocks=1 00:06:15.317 00:06:15.317 ' 00:06:15.317 15:10:43 event.event_scheduler -- scheduler/scheduler.sh@29 -- # rpc=rpc_cmd 00:06:15.317 15:10:43 event.event_scheduler -- scheduler/scheduler.sh@34 -- # /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler -m 0xF -p 0x2 --wait-for-rpc -f 00:06:15.317 15:10:43 event.event_scheduler -- scheduler/scheduler.sh@35 -- # scheduler_pid=70422 00:06:15.317 15:10:43 event.event_scheduler -- scheduler/scheduler.sh@36 -- # trap 'killprocess $scheduler_pid; exit 1' SIGINT SIGTERM EXIT 00:06:15.317 15:10:43 event.event_scheduler -- scheduler/scheduler.sh@37 -- # waitforlisten 70422 00:06:15.317 15:10:43 event.event_scheduler -- common/autotest_common.sh@835 -- # '[' -z 70422 ']' 00:06:15.317 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:15.317 15:10:43 event.event_scheduler -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:15.317 15:10:43 event.event_scheduler -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:15.317 15:10:43 event.event_scheduler -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:15.317 15:10:43 event.event_scheduler -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:15.317 15:10:43 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:06:15.317 [2024-11-27 15:10:43.392591] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:06:15.317 [2024-11-27 15:10:43.392749] [ DPDK EAL parameters: scheduler --no-shconf -c 0xF --main-lcore=2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70422 ] 00:06:15.576 [2024-11-27 15:10:43.563640] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:06:15.576 [2024-11-27 15:10:43.594023] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:15.576 [2024-11-27 15:10:43.594255] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:15.576 [2024-11-27 15:10:43.594285] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:06:15.576 [2024-11-27 15:10:43.594410] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:06:16.144 15:10:44 event.event_scheduler -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:16.144 15:10:44 event.event_scheduler -- common/autotest_common.sh@868 -- # return 0 00:06:16.144 15:10:44 event.event_scheduler -- scheduler/scheduler.sh@39 -- # rpc_cmd framework_set_scheduler dynamic 00:06:16.144 15:10:44 event.event_scheduler -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:16.144 15:10:44 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:06:16.144 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:06:16.144 POWER: Cannot set governor of lcore 0 to userspace 00:06:16.144 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:06:16.144 POWER: Cannot set governor of lcore 0 to performance 00:06:16.144 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:06:16.144 POWER: Cannot set governor of lcore 0 to userspace 00:06:16.144 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:06:16.144 POWER: Cannot set governor of lcore 0 to userspace 00:06:16.144 GUEST_CHANNEL: Unable to connect to '/dev/virtio-ports/virtio.serial.port.poweragent.0' with error No such file or directory 00:06:16.144 POWER: Unable to set Power Management Environment for lcore 0 00:06:16.144 [2024-11-27 15:10:44.235023] dpdk_governor.c: 135:_init_core: *ERROR*: Failed to initialize on core0 00:06:16.144 [2024-11-27 15:10:44.235057] dpdk_governor.c: 196:_init: *ERROR*: Failed to initialize on core0 00:06:16.144 [2024-11-27 15:10:44.235068] scheduler_dynamic.c: 280:init: *NOTICE*: Unable to initialize dpdk governor 00:06:16.144 [2024-11-27 15:10:44.235086] scheduler_dynamic.c: 427:set_opts: *NOTICE*: Setting scheduler load limit to 20 00:06:16.144 [2024-11-27 15:10:44.235096] scheduler_dynamic.c: 429:set_opts: *NOTICE*: Setting scheduler core limit to 80 00:06:16.144 [2024-11-27 15:10:44.235105] scheduler_dynamic.c: 431:set_opts: *NOTICE*: Setting scheduler core busy to 95 00:06:16.144 15:10:44 event.event_scheduler -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:16.144 15:10:44 event.event_scheduler -- scheduler/scheduler.sh@40 -- # rpc_cmd framework_start_init 00:06:16.144 15:10:44 event.event_scheduler -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:16.144 15:10:44 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:06:16.401 [2024-11-27 15:10:44.310724] scheduler.c: 382:test_start: *NOTICE*: Scheduler test application started. 00:06:16.401 15:10:44 event.event_scheduler -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:16.401 15:10:44 event.event_scheduler -- scheduler/scheduler.sh@43 -- # run_test scheduler_create_thread scheduler_create_thread 00:06:16.401 15:10:44 event.event_scheduler -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:16.401 15:10:44 event.event_scheduler -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:16.401 15:10:44 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:06:16.401 ************************************ 00:06:16.401 START TEST scheduler_create_thread 00:06:16.401 ************************************ 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1129 -- # scheduler_create_thread 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@12 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x1 -a 100 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:16.401 2 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@13 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x2 -a 100 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:16.401 3 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@14 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x4 -a 100 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:16.401 4 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@15 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x8 -a 100 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:16.401 5 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@16 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x1 -a 0 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:16.401 6 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@17 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x2 -a 0 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:16.401 7 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@18 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x4 -a 0 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:16.401 8 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@19 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x8 -a 0 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:16.401 9 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@21 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n one_third_active -a 30 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:16.401 10 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:16.401 15:10:44 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n half_active -a 0 00:06:16.402 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:16.402 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:16.402 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:16.402 15:10:44 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # thread_id=11 00:06:16.402 15:10:44 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@23 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_set_active 11 50 00:06:16.402 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:16.402 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:16.402 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:16.402 15:10:44 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n deleted -a 100 00:06:16.402 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:16.402 15:10:44 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:17.776 15:10:45 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:17.776 15:10:45 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # thread_id=12 00:06:17.776 15:10:45 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@26 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_delete 12 00:06:17.776 15:10:45 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:17.776 15:10:45 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:19.151 15:10:46 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:19.151 ************************************ 00:06:19.151 END TEST scheduler_create_thread 00:06:19.151 ************************************ 00:06:19.151 00:06:19.151 real 0m2.610s 00:06:19.151 user 0m0.025s 00:06:19.151 sys 0m0.011s 00:06:19.151 15:10:46 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:19.151 15:10:46 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:19.151 15:10:46 event.event_scheduler -- scheduler/scheduler.sh@45 -- # trap - SIGINT SIGTERM EXIT 00:06:19.151 15:10:46 event.event_scheduler -- scheduler/scheduler.sh@46 -- # killprocess 70422 00:06:19.151 15:10:46 event.event_scheduler -- common/autotest_common.sh@954 -- # '[' -z 70422 ']' 00:06:19.151 15:10:46 event.event_scheduler -- common/autotest_common.sh@958 -- # kill -0 70422 00:06:19.151 15:10:46 event.event_scheduler -- common/autotest_common.sh@959 -- # uname 00:06:19.151 15:10:46 event.event_scheduler -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:19.151 15:10:46 event.event_scheduler -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70422 00:06:19.151 killing process with pid 70422 00:06:19.151 15:10:47 event.event_scheduler -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:06:19.151 15:10:47 event.event_scheduler -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:06:19.151 15:10:47 event.event_scheduler -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70422' 00:06:19.151 15:10:47 event.event_scheduler -- common/autotest_common.sh@973 -- # kill 70422 00:06:19.151 15:10:47 event.event_scheduler -- common/autotest_common.sh@978 -- # wait 70422 00:06:19.410 [2024-11-27 15:10:47.412673] scheduler.c: 360:test_shutdown: *NOTICE*: Scheduler test application stopped. 00:06:19.670 00:06:19.670 real 0m4.553s 00:06:19.670 user 0m8.155s 00:06:19.670 sys 0m0.470s 00:06:19.670 ************************************ 00:06:19.670 END TEST event_scheduler 00:06:19.670 ************************************ 00:06:19.670 15:10:47 event.event_scheduler -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:19.670 15:10:47 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:06:19.670 15:10:47 event -- event/event.sh@51 -- # modprobe -n nbd 00:06:19.670 15:10:47 event -- event/event.sh@52 -- # run_test app_repeat app_repeat_test 00:06:19.670 15:10:47 event -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:19.670 15:10:47 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:19.670 15:10:47 event -- common/autotest_common.sh@10 -- # set +x 00:06:19.670 ************************************ 00:06:19.670 START TEST app_repeat 00:06:19.670 ************************************ 00:06:19.670 15:10:47 event.app_repeat -- common/autotest_common.sh@1129 -- # app_repeat_test 00:06:19.670 15:10:47 event.app_repeat -- event/event.sh@12 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:19.670 15:10:47 event.app_repeat -- event/event.sh@13 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:19.670 15:10:47 event.app_repeat -- event/event.sh@13 -- # local nbd_list 00:06:19.670 15:10:47 event.app_repeat -- event/event.sh@14 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:19.670 15:10:47 event.app_repeat -- event/event.sh@14 -- # local bdev_list 00:06:19.670 15:10:47 event.app_repeat -- event/event.sh@15 -- # local repeat_times=4 00:06:19.670 15:10:47 event.app_repeat -- event/event.sh@17 -- # modprobe nbd 00:06:19.670 15:10:47 event.app_repeat -- event/event.sh@19 -- # repeat_pid=70517 00:06:19.670 15:10:47 event.app_repeat -- event/event.sh@18 -- # /home/vagrant/spdk_repo/spdk/test/event/app_repeat/app_repeat -r /var/tmp/spdk-nbd.sock -m 0x3 -t 4 00:06:19.670 15:10:47 event.app_repeat -- event/event.sh@20 -- # trap 'killprocess $repeat_pid; exit 1' SIGINT SIGTERM EXIT 00:06:19.670 Process app_repeat pid: 70517 00:06:19.670 spdk_app_start Round 0 00:06:19.670 15:10:47 event.app_repeat -- event/event.sh@21 -- # echo 'Process app_repeat pid: 70517' 00:06:19.670 15:10:47 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:06:19.670 15:10:47 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 0' 00:06:19.670 15:10:47 event.app_repeat -- event/event.sh@25 -- # waitforlisten 70517 /var/tmp/spdk-nbd.sock 00:06:19.670 15:10:47 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 70517 ']' 00:06:19.670 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:06:19.670 15:10:47 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:06:19.670 15:10:47 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:19.670 15:10:47 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:06:19.670 15:10:47 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:19.670 15:10:47 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:19.670 [2024-11-27 15:10:47.773594] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:06:19.930 [2024-11-27 15:10:47.773762] [ DPDK EAL parameters: app_repeat --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70517 ] 00:06:19.930 [2024-11-27 15:10:47.943184] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:19.930 [2024-11-27 15:10:47.971270] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:19.930 [2024-11-27 15:10:47.971360] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:20.628 15:10:48 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:20.628 15:10:48 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:06:20.628 15:10:48 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:20.887 Malloc0 00:06:20.887 15:10:48 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:21.156 Malloc1 00:06:21.156 15:10:49 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:21.156 15:10:49 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:21.156 15:10:49 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:21.156 15:10:49 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:06:21.156 15:10:49 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:21.156 15:10:49 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:06:21.156 15:10:49 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:21.156 15:10:49 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:21.156 15:10:49 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:21.156 15:10:49 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:06:21.156 15:10:49 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:21.156 15:10:49 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:06:21.156 15:10:49 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:06:21.156 15:10:49 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:06:21.156 15:10:49 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:21.156 15:10:49 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:06:21.421 /dev/nbd0 00:06:21.421 15:10:49 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:06:21.421 15:10:49 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:06:21.421 15:10:49 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:06:21.421 15:10:49 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:06:21.421 15:10:49 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:06:21.421 15:10:49 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:06:21.421 15:10:49 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:06:21.421 15:10:49 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:06:21.421 15:10:49 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:06:21.421 15:10:49 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:06:21.421 15:10:49 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:21.421 1+0 records in 00:06:21.421 1+0 records out 00:06:21.421 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000176754 s, 23.2 MB/s 00:06:21.421 15:10:49 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:21.421 15:10:49 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:06:21.421 15:10:49 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:21.421 15:10:49 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:06:21.421 15:10:49 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:06:21.421 15:10:49 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:21.421 15:10:49 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:21.421 15:10:49 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:06:21.681 /dev/nbd1 00:06:21.681 15:10:49 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:06:21.681 15:10:49 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:06:21.681 15:10:49 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:06:21.681 15:10:49 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:06:21.681 15:10:49 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:06:21.681 15:10:49 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:06:21.681 15:10:49 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:06:21.681 15:10:49 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:06:21.681 15:10:49 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:06:21.681 15:10:49 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:06:21.681 15:10:49 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:21.681 1+0 records in 00:06:21.681 1+0 records out 00:06:21.681 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000454597 s, 9.0 MB/s 00:06:21.681 15:10:49 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:21.681 15:10:49 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:06:21.681 15:10:49 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:21.681 15:10:49 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:06:21.681 15:10:49 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:06:21.681 15:10:49 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:21.681 15:10:49 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:21.681 15:10:49 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:21.681 15:10:49 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:21.681 15:10:49 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:06:21.941 { 00:06:21.941 "nbd_device": "/dev/nbd0", 00:06:21.941 "bdev_name": "Malloc0" 00:06:21.941 }, 00:06:21.941 { 00:06:21.941 "nbd_device": "/dev/nbd1", 00:06:21.941 "bdev_name": "Malloc1" 00:06:21.941 } 00:06:21.941 ]' 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:06:21.941 { 00:06:21.941 "nbd_device": "/dev/nbd0", 00:06:21.941 "bdev_name": "Malloc0" 00:06:21.941 }, 00:06:21.941 { 00:06:21.941 "nbd_device": "/dev/nbd1", 00:06:21.941 "bdev_name": "Malloc1" 00:06:21.941 } 00:06:21.941 ]' 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:06:21.941 /dev/nbd1' 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:06:21.941 /dev/nbd1' 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:06:21.941 256+0 records in 00:06:21.941 256+0 records out 00:06:21.941 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0136583 s, 76.8 MB/s 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:06:21.941 256+0 records in 00:06:21.941 256+0 records out 00:06:21.941 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0205231 s, 51.1 MB/s 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:06:21.941 256+0 records in 00:06:21.941 256+0 records out 00:06:21.941 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.026393 s, 39.7 MB/s 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:21.941 15:10:49 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:06:22.204 15:10:50 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:06:22.204 15:10:50 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:06:22.204 15:10:50 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:06:22.204 15:10:50 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:22.204 15:10:50 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:22.204 15:10:50 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:06:22.204 15:10:50 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:22.204 15:10:50 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:22.204 15:10:50 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:22.204 15:10:50 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:06:22.463 15:10:50 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:06:22.463 15:10:50 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:06:22.463 15:10:50 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:06:22.463 15:10:50 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:22.463 15:10:50 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:22.463 15:10:50 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:06:22.463 15:10:50 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:22.463 15:10:50 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:22.463 15:10:50 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:22.463 15:10:50 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:22.463 15:10:50 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:22.722 15:10:50 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:06:22.722 15:10:50 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:06:22.722 15:10:50 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:22.722 15:10:50 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:06:22.722 15:10:50 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:06:22.722 15:10:50 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:22.723 15:10:50 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:06:22.723 15:10:50 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:06:22.723 15:10:50 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:06:22.723 15:10:50 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:06:22.723 15:10:50 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:06:22.723 15:10:50 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:06:22.723 15:10:50 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:06:22.982 15:10:50 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:06:22.982 [2024-11-27 15:10:50.976130] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:22.982 [2024-11-27 15:10:50.999973] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:22.982 [2024-11-27 15:10:50.999977] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:22.982 [2024-11-27 15:10:51.043102] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:06:22.982 [2024-11-27 15:10:51.043283] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:06:26.278 15:10:53 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:06:26.278 15:10:53 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 1' 00:06:26.278 spdk_app_start Round 1 00:06:26.278 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:06:26.278 15:10:53 event.app_repeat -- event/event.sh@25 -- # waitforlisten 70517 /var/tmp/spdk-nbd.sock 00:06:26.278 15:10:53 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 70517 ']' 00:06:26.278 15:10:53 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:06:26.278 15:10:53 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:26.278 15:10:53 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:06:26.278 15:10:53 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:26.278 15:10:53 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:26.278 15:10:54 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:26.278 15:10:54 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:06:26.278 15:10:54 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:26.278 Malloc0 00:06:26.278 15:10:54 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:26.538 Malloc1 00:06:26.538 15:10:54 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:26.538 15:10:54 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:26.538 15:10:54 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:26.538 15:10:54 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:06:26.538 15:10:54 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:26.538 15:10:54 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:06:26.538 15:10:54 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:26.538 15:10:54 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:26.538 15:10:54 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:26.538 15:10:54 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:06:26.538 15:10:54 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:26.538 15:10:54 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:06:26.538 15:10:54 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:06:26.538 15:10:54 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:06:26.538 15:10:54 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:26.538 15:10:54 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:06:26.798 /dev/nbd0 00:06:26.798 15:10:54 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:06:26.798 15:10:54 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:06:26.798 15:10:54 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:06:26.798 15:10:54 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:06:26.798 15:10:54 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:06:26.798 15:10:54 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:06:26.798 15:10:54 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:06:26.798 15:10:54 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:06:26.798 15:10:54 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:06:26.798 15:10:54 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:06:26.798 15:10:54 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:26.798 1+0 records in 00:06:26.798 1+0 records out 00:06:26.798 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000474737 s, 8.6 MB/s 00:06:26.798 15:10:54 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:26.798 15:10:54 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:06:26.798 15:10:54 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:26.798 15:10:54 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:06:26.798 15:10:54 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:06:26.798 15:10:54 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:26.798 15:10:54 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:26.798 15:10:54 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:06:27.059 /dev/nbd1 00:06:27.059 15:10:54 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:06:27.059 15:10:54 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:06:27.059 15:10:54 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:06:27.059 15:10:54 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:06:27.059 15:10:54 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:06:27.059 15:10:54 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:06:27.059 15:10:54 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:06:27.059 15:10:54 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:06:27.059 15:10:54 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:06:27.059 15:10:54 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:06:27.059 15:10:54 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:27.059 1+0 records in 00:06:27.059 1+0 records out 00:06:27.059 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000486831 s, 8.4 MB/s 00:06:27.059 15:10:54 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:27.059 15:10:54 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:06:27.059 15:10:54 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:27.059 15:10:54 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:06:27.059 15:10:54 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:06:27.059 15:10:54 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:27.059 15:10:54 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:27.059 15:10:54 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:27.059 15:10:54 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:27.059 15:10:54 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:06:27.319 { 00:06:27.319 "nbd_device": "/dev/nbd0", 00:06:27.319 "bdev_name": "Malloc0" 00:06:27.319 }, 00:06:27.319 { 00:06:27.319 "nbd_device": "/dev/nbd1", 00:06:27.319 "bdev_name": "Malloc1" 00:06:27.319 } 00:06:27.319 ]' 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:06:27.319 { 00:06:27.319 "nbd_device": "/dev/nbd0", 00:06:27.319 "bdev_name": "Malloc0" 00:06:27.319 }, 00:06:27.319 { 00:06:27.319 "nbd_device": "/dev/nbd1", 00:06:27.319 "bdev_name": "Malloc1" 00:06:27.319 } 00:06:27.319 ]' 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:06:27.319 /dev/nbd1' 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:06:27.319 /dev/nbd1' 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:06:27.319 256+0 records in 00:06:27.319 256+0 records out 00:06:27.319 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00434097 s, 242 MB/s 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:06:27.319 256+0 records in 00:06:27.319 256+0 records out 00:06:27.319 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0249099 s, 42.1 MB/s 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:06:27.319 256+0 records in 00:06:27.319 256+0 records out 00:06:27.319 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0243686 s, 43.0 MB/s 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:27.319 15:10:55 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:06:27.578 15:10:55 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:06:27.578 15:10:55 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:06:27.578 15:10:55 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:06:27.578 15:10:55 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:27.578 15:10:55 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:27.578 15:10:55 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:06:27.578 15:10:55 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:27.578 15:10:55 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:27.578 15:10:55 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:27.578 15:10:55 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:06:27.838 15:10:55 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:06:27.838 15:10:55 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:06:27.838 15:10:55 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:06:27.838 15:10:55 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:27.838 15:10:55 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:27.838 15:10:55 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:06:27.838 15:10:55 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:27.838 15:10:55 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:27.838 15:10:55 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:27.838 15:10:55 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:27.838 15:10:55 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:27.838 15:10:55 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:06:27.838 15:10:55 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:06:27.839 15:10:55 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:28.098 15:10:55 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:06:28.098 15:10:55 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:06:28.098 15:10:55 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:28.098 15:10:55 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:06:28.098 15:10:55 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:06:28.098 15:10:55 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:06:28.098 15:10:55 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:06:28.098 15:10:55 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:06:28.098 15:10:55 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:06:28.098 15:10:55 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:06:28.357 15:10:56 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:06:28.357 [2024-11-27 15:10:56.347503] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:28.357 [2024-11-27 15:10:56.373336] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:28.357 [2024-11-27 15:10:56.373364] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:28.357 [2024-11-27 15:10:56.416785] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:06:28.357 [2024-11-27 15:10:56.416982] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:06:31.649 spdk_app_start Round 2 00:06:31.649 15:10:59 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:06:31.649 15:10:59 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 2' 00:06:31.649 15:10:59 event.app_repeat -- event/event.sh@25 -- # waitforlisten 70517 /var/tmp/spdk-nbd.sock 00:06:31.649 15:10:59 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 70517 ']' 00:06:31.649 15:10:59 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:06:31.649 15:10:59 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:31.650 15:10:59 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:06:31.650 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:06:31.650 15:10:59 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:31.650 15:10:59 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:31.650 15:10:59 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:31.650 15:10:59 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:06:31.650 15:10:59 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:31.650 Malloc0 00:06:31.650 15:10:59 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:31.909 Malloc1 00:06:31.909 15:10:59 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:31.909 15:10:59 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:31.909 15:10:59 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:31.909 15:10:59 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:06:31.909 15:10:59 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:31.909 15:10:59 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:06:31.909 15:10:59 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:31.909 15:10:59 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:31.909 15:10:59 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:31.909 15:10:59 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:06:31.909 15:10:59 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:31.909 15:10:59 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:06:31.909 15:10:59 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:06:31.909 15:10:59 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:06:31.909 15:10:59 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:31.909 15:10:59 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:06:32.167 /dev/nbd0 00:06:32.167 15:11:00 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:06:32.167 15:11:00 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:06:32.167 15:11:00 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:06:32.167 15:11:00 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:06:32.167 15:11:00 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:06:32.167 15:11:00 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:06:32.167 15:11:00 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:06:32.167 15:11:00 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:06:32.167 15:11:00 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:06:32.167 15:11:00 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:06:32.167 15:11:00 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:32.167 1+0 records in 00:06:32.167 1+0 records out 00:06:32.167 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000659083 s, 6.2 MB/s 00:06:32.167 15:11:00 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:32.167 15:11:00 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:06:32.167 15:11:00 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:32.167 15:11:00 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:06:32.167 15:11:00 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:06:32.167 15:11:00 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:32.167 15:11:00 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:32.167 15:11:00 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:06:32.427 /dev/nbd1 00:06:32.427 15:11:00 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:06:32.427 15:11:00 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:06:32.427 15:11:00 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:06:32.427 15:11:00 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:06:32.427 15:11:00 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:06:32.427 15:11:00 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:06:32.427 15:11:00 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:06:32.427 15:11:00 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:06:32.427 15:11:00 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:06:32.427 15:11:00 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:06:32.427 15:11:00 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:32.427 1+0 records in 00:06:32.427 1+0 records out 00:06:32.427 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000296117 s, 13.8 MB/s 00:06:32.427 15:11:00 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:32.427 15:11:00 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:06:32.427 15:11:00 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:32.427 15:11:00 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:06:32.427 15:11:00 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:06:32.427 15:11:00 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:32.427 15:11:00 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:32.427 15:11:00 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:32.427 15:11:00 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:32.427 15:11:00 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:06:32.688 { 00:06:32.688 "nbd_device": "/dev/nbd0", 00:06:32.688 "bdev_name": "Malloc0" 00:06:32.688 }, 00:06:32.688 { 00:06:32.688 "nbd_device": "/dev/nbd1", 00:06:32.688 "bdev_name": "Malloc1" 00:06:32.688 } 00:06:32.688 ]' 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:06:32.688 { 00:06:32.688 "nbd_device": "/dev/nbd0", 00:06:32.688 "bdev_name": "Malloc0" 00:06:32.688 }, 00:06:32.688 { 00:06:32.688 "nbd_device": "/dev/nbd1", 00:06:32.688 "bdev_name": "Malloc1" 00:06:32.688 } 00:06:32.688 ]' 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:06:32.688 /dev/nbd1' 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:06:32.688 /dev/nbd1' 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:06:32.688 256+0 records in 00:06:32.688 256+0 records out 00:06:32.688 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0139148 s, 75.4 MB/s 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:06:32.688 256+0 records in 00:06:32.688 256+0 records out 00:06:32.688 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0268763 s, 39.0 MB/s 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:06:32.688 256+0 records in 00:06:32.688 256+0 records out 00:06:32.688 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0261473 s, 40.1 MB/s 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:32.688 15:11:00 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:06:32.689 15:11:00 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:32.689 15:11:00 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:32.689 15:11:00 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:06:32.689 15:11:00 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:06:32.689 15:11:00 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:32.689 15:11:00 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:06:32.949 15:11:00 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:06:32.949 15:11:00 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:06:32.949 15:11:00 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:06:32.949 15:11:00 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:32.949 15:11:00 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:32.949 15:11:00 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:06:32.949 15:11:00 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:32.949 15:11:00 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:32.949 15:11:00 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:32.949 15:11:00 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:06:33.209 15:11:01 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:06:33.209 15:11:01 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:06:33.209 15:11:01 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:06:33.209 15:11:01 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:33.209 15:11:01 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:33.209 15:11:01 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:06:33.209 15:11:01 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:33.209 15:11:01 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:33.209 15:11:01 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:33.209 15:11:01 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:33.209 15:11:01 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:33.469 15:11:01 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:06:33.469 15:11:01 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:33.469 15:11:01 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:06:33.469 15:11:01 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:06:33.469 15:11:01 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:06:33.469 15:11:01 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:33.469 15:11:01 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:06:33.469 15:11:01 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:06:33.469 15:11:01 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:06:33.469 15:11:01 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:06:33.469 15:11:01 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:06:33.469 15:11:01 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:06:33.469 15:11:01 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:06:33.729 15:11:01 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:06:33.729 [2024-11-27 15:11:01.738763] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:33.729 [2024-11-27 15:11:01.763758] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:33.729 [2024-11-27 15:11:01.763761] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:33.729 [2024-11-27 15:11:01.807613] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:06:33.729 [2024-11-27 15:11:01.807673] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:06:37.023 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:06:37.023 15:11:04 event.app_repeat -- event/event.sh@38 -- # waitforlisten 70517 /var/tmp/spdk-nbd.sock 00:06:37.023 15:11:04 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 70517 ']' 00:06:37.023 15:11:04 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:06:37.023 15:11:04 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:37.023 15:11:04 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:06:37.023 15:11:04 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:37.023 15:11:04 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:37.023 15:11:04 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:37.023 15:11:04 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:06:37.023 15:11:04 event.app_repeat -- event/event.sh@39 -- # killprocess 70517 00:06:37.023 15:11:04 event.app_repeat -- common/autotest_common.sh@954 -- # '[' -z 70517 ']' 00:06:37.023 15:11:04 event.app_repeat -- common/autotest_common.sh@958 -- # kill -0 70517 00:06:37.023 15:11:04 event.app_repeat -- common/autotest_common.sh@959 -- # uname 00:06:37.023 15:11:04 event.app_repeat -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:37.023 15:11:04 event.app_repeat -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70517 00:06:37.023 15:11:04 event.app_repeat -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:37.023 15:11:04 event.app_repeat -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:37.023 15:11:04 event.app_repeat -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70517' 00:06:37.023 killing process with pid 70517 00:06:37.023 15:11:04 event.app_repeat -- common/autotest_common.sh@973 -- # kill 70517 00:06:37.023 15:11:04 event.app_repeat -- common/autotest_common.sh@978 -- # wait 70517 00:06:37.023 spdk_app_start is called in Round 0. 00:06:37.023 Shutdown signal received, stop current app iteration 00:06:37.023 Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 reinitialization... 00:06:37.023 spdk_app_start is called in Round 1. 00:06:37.023 Shutdown signal received, stop current app iteration 00:06:37.023 Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 reinitialization... 00:06:37.023 spdk_app_start is called in Round 2. 00:06:37.023 Shutdown signal received, stop current app iteration 00:06:37.023 Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 reinitialization... 00:06:37.023 spdk_app_start is called in Round 3. 00:06:37.023 Shutdown signal received, stop current app iteration 00:06:37.023 15:11:05 event.app_repeat -- event/event.sh@40 -- # trap - SIGINT SIGTERM EXIT 00:06:37.023 15:11:05 event.app_repeat -- event/event.sh@42 -- # return 0 00:06:37.023 00:06:37.023 real 0m17.315s 00:06:37.023 user 0m38.509s 00:06:37.023 sys 0m2.316s 00:06:37.023 15:11:05 event.app_repeat -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:37.023 15:11:05 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:37.023 ************************************ 00:06:37.023 END TEST app_repeat 00:06:37.023 ************************************ 00:06:37.023 15:11:05 event -- event/event.sh@54 -- # (( SPDK_TEST_CRYPTO == 0 )) 00:06:37.023 15:11:05 event -- event/event.sh@55 -- # run_test cpu_locks /home/vagrant/spdk_repo/spdk/test/event/cpu_locks.sh 00:06:37.023 15:11:05 event -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:37.023 15:11:05 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:37.023 15:11:05 event -- common/autotest_common.sh@10 -- # set +x 00:06:37.023 ************************************ 00:06:37.023 START TEST cpu_locks 00:06:37.023 ************************************ 00:06:37.023 15:11:05 event.cpu_locks -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/cpu_locks.sh 00:06:37.332 * Looking for test storage... 00:06:37.332 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event 00:06:37.332 15:11:05 event.cpu_locks -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:06:37.332 15:11:05 event.cpu_locks -- common/autotest_common.sh@1693 -- # lcov --version 00:06:37.332 15:11:05 event.cpu_locks -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:06:37.332 15:11:05 event.cpu_locks -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:06:37.332 15:11:05 event.cpu_locks -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:37.332 15:11:05 event.cpu_locks -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:37.332 15:11:05 event.cpu_locks -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:37.332 15:11:05 event.cpu_locks -- scripts/common.sh@336 -- # IFS=.-: 00:06:37.332 15:11:05 event.cpu_locks -- scripts/common.sh@336 -- # read -ra ver1 00:06:37.332 15:11:05 event.cpu_locks -- scripts/common.sh@337 -- # IFS=.-: 00:06:37.332 15:11:05 event.cpu_locks -- scripts/common.sh@337 -- # read -ra ver2 00:06:37.332 15:11:05 event.cpu_locks -- scripts/common.sh@338 -- # local 'op=<' 00:06:37.332 15:11:05 event.cpu_locks -- scripts/common.sh@340 -- # ver1_l=2 00:06:37.332 15:11:05 event.cpu_locks -- scripts/common.sh@341 -- # ver2_l=1 00:06:37.332 15:11:05 event.cpu_locks -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:37.332 15:11:05 event.cpu_locks -- scripts/common.sh@344 -- # case "$op" in 00:06:37.332 15:11:05 event.cpu_locks -- scripts/common.sh@345 -- # : 1 00:06:37.332 15:11:05 event.cpu_locks -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:37.332 15:11:05 event.cpu_locks -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:37.332 15:11:05 event.cpu_locks -- scripts/common.sh@365 -- # decimal 1 00:06:37.332 15:11:05 event.cpu_locks -- scripts/common.sh@353 -- # local d=1 00:06:37.332 15:11:05 event.cpu_locks -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:37.332 15:11:05 event.cpu_locks -- scripts/common.sh@355 -- # echo 1 00:06:37.332 15:11:05 event.cpu_locks -- scripts/common.sh@365 -- # ver1[v]=1 00:06:37.332 15:11:05 event.cpu_locks -- scripts/common.sh@366 -- # decimal 2 00:06:37.332 15:11:05 event.cpu_locks -- scripts/common.sh@353 -- # local d=2 00:06:37.332 15:11:05 event.cpu_locks -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:37.332 15:11:05 event.cpu_locks -- scripts/common.sh@355 -- # echo 2 00:06:37.332 15:11:05 event.cpu_locks -- scripts/common.sh@366 -- # ver2[v]=2 00:06:37.332 15:11:05 event.cpu_locks -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:37.332 15:11:05 event.cpu_locks -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:37.332 15:11:05 event.cpu_locks -- scripts/common.sh@368 -- # return 0 00:06:37.332 15:11:05 event.cpu_locks -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:37.332 15:11:05 event.cpu_locks -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:06:37.332 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:37.332 --rc genhtml_branch_coverage=1 00:06:37.332 --rc genhtml_function_coverage=1 00:06:37.332 --rc genhtml_legend=1 00:06:37.332 --rc geninfo_all_blocks=1 00:06:37.332 --rc geninfo_unexecuted_blocks=1 00:06:37.332 00:06:37.332 ' 00:06:37.332 15:11:05 event.cpu_locks -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:06:37.332 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:37.332 --rc genhtml_branch_coverage=1 00:06:37.332 --rc genhtml_function_coverage=1 00:06:37.332 --rc genhtml_legend=1 00:06:37.332 --rc geninfo_all_blocks=1 00:06:37.332 --rc geninfo_unexecuted_blocks=1 00:06:37.332 00:06:37.332 ' 00:06:37.332 15:11:05 event.cpu_locks -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:06:37.332 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:37.332 --rc genhtml_branch_coverage=1 00:06:37.332 --rc genhtml_function_coverage=1 00:06:37.333 --rc genhtml_legend=1 00:06:37.333 --rc geninfo_all_blocks=1 00:06:37.333 --rc geninfo_unexecuted_blocks=1 00:06:37.333 00:06:37.333 ' 00:06:37.333 15:11:05 event.cpu_locks -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:06:37.333 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:37.333 --rc genhtml_branch_coverage=1 00:06:37.333 --rc genhtml_function_coverage=1 00:06:37.333 --rc genhtml_legend=1 00:06:37.333 --rc geninfo_all_blocks=1 00:06:37.333 --rc geninfo_unexecuted_blocks=1 00:06:37.333 00:06:37.333 ' 00:06:37.333 15:11:05 event.cpu_locks -- event/cpu_locks.sh@11 -- # rpc_sock1=/var/tmp/spdk.sock 00:06:37.333 15:11:05 event.cpu_locks -- event/cpu_locks.sh@12 -- # rpc_sock2=/var/tmp/spdk2.sock 00:06:37.333 15:11:05 event.cpu_locks -- event/cpu_locks.sh@164 -- # trap cleanup EXIT SIGTERM SIGINT 00:06:37.333 15:11:05 event.cpu_locks -- event/cpu_locks.sh@166 -- # run_test default_locks default_locks 00:06:37.333 15:11:05 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:37.333 15:11:05 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:37.333 15:11:05 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:37.333 ************************************ 00:06:37.333 START TEST default_locks 00:06:37.333 ************************************ 00:06:37.333 15:11:05 event.cpu_locks.default_locks -- common/autotest_common.sh@1129 -- # default_locks 00:06:37.333 15:11:05 event.cpu_locks.default_locks -- event/cpu_locks.sh@46 -- # spdk_tgt_pid=70943 00:06:37.333 15:11:05 event.cpu_locks.default_locks -- event/cpu_locks.sh@47 -- # waitforlisten 70943 00:06:37.333 15:11:05 event.cpu_locks.default_locks -- common/autotest_common.sh@835 -- # '[' -z 70943 ']' 00:06:37.333 15:11:05 event.cpu_locks.default_locks -- event/cpu_locks.sh@45 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:37.333 15:11:05 event.cpu_locks.default_locks -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:37.333 15:11:05 event.cpu_locks.default_locks -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:37.333 15:11:05 event.cpu_locks.default_locks -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:37.333 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:37.333 15:11:05 event.cpu_locks.default_locks -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:37.333 15:11:05 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:06:37.333 [2024-11-27 15:11:05.432499] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:06:37.333 [2024-11-27 15:11:05.433084] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70943 ] 00:06:37.595 [2024-11-27 15:11:05.604131] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:37.595 [2024-11-27 15:11:05.631604] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:38.165 15:11:06 event.cpu_locks.default_locks -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:38.165 15:11:06 event.cpu_locks.default_locks -- common/autotest_common.sh@868 -- # return 0 00:06:38.165 15:11:06 event.cpu_locks.default_locks -- event/cpu_locks.sh@49 -- # locks_exist 70943 00:06:38.165 15:11:06 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # lslocks -p 70943 00:06:38.165 15:11:06 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:06:38.733 15:11:06 event.cpu_locks.default_locks -- event/cpu_locks.sh@50 -- # killprocess 70943 00:06:38.733 15:11:06 event.cpu_locks.default_locks -- common/autotest_common.sh@954 -- # '[' -z 70943 ']' 00:06:38.733 15:11:06 event.cpu_locks.default_locks -- common/autotest_common.sh@958 -- # kill -0 70943 00:06:38.733 15:11:06 event.cpu_locks.default_locks -- common/autotest_common.sh@959 -- # uname 00:06:38.733 15:11:06 event.cpu_locks.default_locks -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:38.733 15:11:06 event.cpu_locks.default_locks -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70943 00:06:38.733 killing process with pid 70943 00:06:38.733 15:11:06 event.cpu_locks.default_locks -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:38.733 15:11:06 event.cpu_locks.default_locks -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:38.733 15:11:06 event.cpu_locks.default_locks -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70943' 00:06:38.733 15:11:06 event.cpu_locks.default_locks -- common/autotest_common.sh@973 -- # kill 70943 00:06:38.734 15:11:06 event.cpu_locks.default_locks -- common/autotest_common.sh@978 -- # wait 70943 00:06:38.994 15:11:07 event.cpu_locks.default_locks -- event/cpu_locks.sh@52 -- # NOT waitforlisten 70943 00:06:38.994 15:11:07 event.cpu_locks.default_locks -- common/autotest_common.sh@652 -- # local es=0 00:06:38.994 15:11:07 event.cpu_locks.default_locks -- common/autotest_common.sh@654 -- # valid_exec_arg waitforlisten 70943 00:06:38.994 15:11:07 event.cpu_locks.default_locks -- common/autotest_common.sh@640 -- # local arg=waitforlisten 00:06:38.994 15:11:07 event.cpu_locks.default_locks -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:38.994 15:11:07 event.cpu_locks.default_locks -- common/autotest_common.sh@644 -- # type -t waitforlisten 00:06:38.994 15:11:07 event.cpu_locks.default_locks -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:38.994 15:11:07 event.cpu_locks.default_locks -- common/autotest_common.sh@655 -- # waitforlisten 70943 00:06:38.994 15:11:07 event.cpu_locks.default_locks -- common/autotest_common.sh@835 -- # '[' -z 70943 ']' 00:06:38.994 15:11:07 event.cpu_locks.default_locks -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:38.994 15:11:07 event.cpu_locks.default_locks -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:38.994 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:38.994 ERROR: process (pid: 70943) is no longer running 00:06:38.994 15:11:07 event.cpu_locks.default_locks -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:38.994 15:11:07 event.cpu_locks.default_locks -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:38.994 15:11:07 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:06:38.994 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 850: kill: (70943) - No such process 00:06:38.994 15:11:07 event.cpu_locks.default_locks -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:38.994 15:11:07 event.cpu_locks.default_locks -- common/autotest_common.sh@868 -- # return 1 00:06:38.994 15:11:07 event.cpu_locks.default_locks -- common/autotest_common.sh@655 -- # es=1 00:06:38.994 15:11:07 event.cpu_locks.default_locks -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:06:38.994 15:11:07 event.cpu_locks.default_locks -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:06:38.994 15:11:07 event.cpu_locks.default_locks -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:06:38.994 15:11:07 event.cpu_locks.default_locks -- event/cpu_locks.sh@54 -- # no_locks 00:06:38.994 15:11:07 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # lock_files=() 00:06:38.994 15:11:07 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # local lock_files 00:06:38.994 15:11:07 event.cpu_locks.default_locks -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:06:38.994 00:06:38.994 real 0m1.743s 00:06:38.994 user 0m1.707s 00:06:38.994 sys 0m0.616s 00:06:38.994 15:11:07 event.cpu_locks.default_locks -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:38.994 15:11:07 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:06:38.994 ************************************ 00:06:38.994 END TEST default_locks 00:06:38.994 ************************************ 00:06:39.255 15:11:07 event.cpu_locks -- event/cpu_locks.sh@167 -- # run_test default_locks_via_rpc default_locks_via_rpc 00:06:39.255 15:11:07 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:39.255 15:11:07 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:39.255 15:11:07 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:39.255 ************************************ 00:06:39.255 START TEST default_locks_via_rpc 00:06:39.255 ************************************ 00:06:39.255 15:11:07 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1129 -- # default_locks_via_rpc 00:06:39.255 15:11:07 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@62 -- # spdk_tgt_pid=70992 00:06:39.255 15:11:07 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@61 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:39.255 15:11:07 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@63 -- # waitforlisten 70992 00:06:39.255 15:11:07 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 70992 ']' 00:06:39.256 15:11:07 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:39.256 15:11:07 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:39.256 15:11:07 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:39.256 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:39.256 15:11:07 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:39.256 15:11:07 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:39.256 [2024-11-27 15:11:07.243351] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:06:39.256 [2024-11-27 15:11:07.243565] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70992 ] 00:06:39.515 [2024-11-27 15:11:07.411408] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:39.515 [2024-11-27 15:11:07.437149] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:40.085 15:11:08 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:40.085 15:11:08 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:06:40.085 15:11:08 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@65 -- # rpc_cmd framework_disable_cpumask_locks 00:06:40.085 15:11:08 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:40.085 15:11:08 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:40.085 15:11:08 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:40.085 15:11:08 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@67 -- # no_locks 00:06:40.085 15:11:08 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # lock_files=() 00:06:40.085 15:11:08 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # local lock_files 00:06:40.085 15:11:08 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:06:40.085 15:11:08 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@69 -- # rpc_cmd framework_enable_cpumask_locks 00:06:40.085 15:11:08 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:40.085 15:11:08 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:40.086 15:11:08 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:40.086 15:11:08 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@71 -- # locks_exist 70992 00:06:40.086 15:11:08 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:06:40.086 15:11:08 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # lslocks -p 70992 00:06:40.656 15:11:08 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@73 -- # killprocess 70992 00:06:40.656 15:11:08 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@954 -- # '[' -z 70992 ']' 00:06:40.656 15:11:08 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@958 -- # kill -0 70992 00:06:40.656 15:11:08 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@959 -- # uname 00:06:40.656 15:11:08 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:40.656 15:11:08 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70992 00:06:40.656 15:11:08 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:40.656 15:11:08 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:40.656 15:11:08 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70992' 00:06:40.656 killing process with pid 70992 00:06:40.656 15:11:08 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@973 -- # kill 70992 00:06:40.656 15:11:08 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@978 -- # wait 70992 00:06:40.916 00:06:40.916 real 0m1.772s 00:06:40.916 user 0m1.745s 00:06:40.916 sys 0m0.617s 00:06:40.916 15:11:08 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:40.916 ************************************ 00:06:40.916 END TEST default_locks_via_rpc 00:06:40.916 ************************************ 00:06:40.916 15:11:08 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:40.916 15:11:08 event.cpu_locks -- event/cpu_locks.sh@168 -- # run_test non_locking_app_on_locked_coremask non_locking_app_on_locked_coremask 00:06:40.916 15:11:08 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:40.916 15:11:08 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:40.916 15:11:08 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:40.916 ************************************ 00:06:40.916 START TEST non_locking_app_on_locked_coremask 00:06:40.916 ************************************ 00:06:40.916 15:11:08 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1129 -- # non_locking_app_on_locked_coremask 00:06:40.916 15:11:08 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@80 -- # spdk_tgt_pid=71044 00:06:40.916 15:11:08 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@79 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:40.916 15:11:08 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@81 -- # waitforlisten 71044 /var/tmp/spdk.sock 00:06:40.916 15:11:08 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 71044 ']' 00:06:40.916 15:11:08 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:40.916 15:11:08 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:40.916 15:11:08 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:40.916 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:40.916 15:11:08 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:40.917 15:11:08 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:41.176 [2024-11-27 15:11:09.087400] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:06:41.176 [2024-11-27 15:11:09.087541] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71044 ] 00:06:41.176 [2024-11-27 15:11:09.259433] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:41.436 [2024-11-27 15:11:09.286021] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:42.006 15:11:09 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:42.006 15:11:09 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 0 00:06:42.006 15:11:09 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@84 -- # spdk_tgt_pid2=71060 00:06:42.006 15:11:09 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@83 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks -r /var/tmp/spdk2.sock 00:06:42.006 15:11:09 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@85 -- # waitforlisten 71060 /var/tmp/spdk2.sock 00:06:42.006 15:11:09 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 71060 ']' 00:06:42.006 15:11:09 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:42.006 15:11:09 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:42.006 15:11:09 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:42.006 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:42.006 15:11:09 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:42.006 15:11:09 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:42.006 [2024-11-27 15:11:09.987724] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:06:42.006 [2024-11-27 15:11:09.987979] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71060 ] 00:06:42.267 [2024-11-27 15:11:10.152217] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:06:42.267 [2024-11-27 15:11:10.152277] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:42.267 [2024-11-27 15:11:10.206628] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:42.836 15:11:10 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:42.836 15:11:10 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 0 00:06:42.836 15:11:10 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@87 -- # locks_exist 71044 00:06:42.836 15:11:10 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 71044 00:06:42.836 15:11:10 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:06:43.097 15:11:11 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@89 -- # killprocess 71044 00:06:43.097 15:11:11 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # '[' -z 71044 ']' 00:06:43.097 15:11:11 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # kill -0 71044 00:06:43.097 15:11:11 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # uname 00:06:43.097 15:11:11 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:43.097 15:11:11 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71044 00:06:43.097 killing process with pid 71044 00:06:43.097 15:11:11 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:43.097 15:11:11 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:43.097 15:11:11 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71044' 00:06:43.097 15:11:11 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@973 -- # kill 71044 00:06:43.097 15:11:11 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@978 -- # wait 71044 00:06:44.037 15:11:11 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@90 -- # killprocess 71060 00:06:44.037 15:11:11 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # '[' -z 71060 ']' 00:06:44.037 15:11:11 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # kill -0 71060 00:06:44.037 15:11:11 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # uname 00:06:44.037 15:11:11 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:44.037 15:11:11 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71060 00:06:44.037 killing process with pid 71060 00:06:44.037 15:11:11 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:44.037 15:11:11 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:44.037 15:11:11 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71060' 00:06:44.037 15:11:11 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@973 -- # kill 71060 00:06:44.037 15:11:11 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@978 -- # wait 71060 00:06:44.297 00:06:44.297 real 0m3.257s 00:06:44.297 user 0m3.456s 00:06:44.297 sys 0m0.949s 00:06:44.297 15:11:12 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:44.297 15:11:12 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:44.297 ************************************ 00:06:44.297 END TEST non_locking_app_on_locked_coremask 00:06:44.297 ************************************ 00:06:44.297 15:11:12 event.cpu_locks -- event/cpu_locks.sh@169 -- # run_test locking_app_on_unlocked_coremask locking_app_on_unlocked_coremask 00:06:44.297 15:11:12 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:44.298 15:11:12 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:44.298 15:11:12 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:44.298 ************************************ 00:06:44.298 START TEST locking_app_on_unlocked_coremask 00:06:44.298 ************************************ 00:06:44.298 15:11:12 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1129 -- # locking_app_on_unlocked_coremask 00:06:44.298 15:11:12 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@98 -- # spdk_tgt_pid=71119 00:06:44.298 15:11:12 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@97 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks 00:06:44.298 15:11:12 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@99 -- # waitforlisten 71119 /var/tmp/spdk.sock 00:06:44.298 15:11:12 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@835 -- # '[' -z 71119 ']' 00:06:44.298 15:11:12 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:44.298 15:11:12 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:44.298 15:11:12 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:44.298 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:44.298 15:11:12 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:44.298 15:11:12 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:44.558 [2024-11-27 15:11:12.411189] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:06:44.558 [2024-11-27 15:11:12.411789] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71119 ] 00:06:44.558 [2024-11-27 15:11:12.580954] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:06:44.558 [2024-11-27 15:11:12.581097] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:44.558 [2024-11-27 15:11:12.606232] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:45.128 15:11:13 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:45.128 15:11:13 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@868 -- # return 0 00:06:45.128 15:11:13 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@101 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:06:45.128 15:11:13 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@102 -- # spdk_tgt_pid2=71135 00:06:45.128 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:45.128 15:11:13 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@103 -- # waitforlisten 71135 /var/tmp/spdk2.sock 00:06:45.128 15:11:13 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@835 -- # '[' -z 71135 ']' 00:06:45.128 15:11:13 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:45.128 15:11:13 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:45.128 15:11:13 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:45.128 15:11:13 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:45.128 15:11:13 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:45.388 [2024-11-27 15:11:13.294803] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:06:45.388 [2024-11-27 15:11:13.294956] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71135 ] 00:06:45.388 [2024-11-27 15:11:13.459401] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:45.648 [2024-11-27 15:11:13.515229] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:46.218 15:11:14 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:46.218 15:11:14 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@868 -- # return 0 00:06:46.218 15:11:14 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@105 -- # locks_exist 71135 00:06:46.218 15:11:14 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 71135 00:06:46.218 15:11:14 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:06:46.786 15:11:14 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@107 -- # killprocess 71119 00:06:46.786 15:11:14 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # '[' -z 71119 ']' 00:06:46.786 15:11:14 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@958 -- # kill -0 71119 00:06:46.786 15:11:14 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # uname 00:06:46.786 15:11:14 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:46.786 15:11:14 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71119 00:06:47.045 15:11:14 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:47.046 killing process with pid 71119 00:06:47.046 15:11:14 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:47.046 15:11:14 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71119' 00:06:47.046 15:11:14 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@973 -- # kill 71119 00:06:47.046 15:11:14 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@978 -- # wait 71119 00:06:47.615 15:11:15 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@108 -- # killprocess 71135 00:06:47.615 15:11:15 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # '[' -z 71135 ']' 00:06:47.616 15:11:15 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@958 -- # kill -0 71135 00:06:47.616 15:11:15 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # uname 00:06:47.616 15:11:15 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:47.616 15:11:15 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71135 00:06:47.616 killing process with pid 71135 00:06:47.616 15:11:15 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:47.616 15:11:15 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:47.616 15:11:15 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71135' 00:06:47.616 15:11:15 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@973 -- # kill 71135 00:06:47.616 15:11:15 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@978 -- # wait 71135 00:06:48.184 00:06:48.184 real 0m3.699s 00:06:48.184 user 0m3.834s 00:06:48.184 sys 0m1.184s 00:06:48.184 ************************************ 00:06:48.184 END TEST locking_app_on_unlocked_coremask 00:06:48.184 ************************************ 00:06:48.184 15:11:16 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:48.184 15:11:16 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:48.184 15:11:16 event.cpu_locks -- event/cpu_locks.sh@170 -- # run_test locking_app_on_locked_coremask locking_app_on_locked_coremask 00:06:48.184 15:11:16 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:48.184 15:11:16 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:48.184 15:11:16 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:48.184 ************************************ 00:06:48.185 START TEST locking_app_on_locked_coremask 00:06:48.185 ************************************ 00:06:48.185 15:11:16 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1129 -- # locking_app_on_locked_coremask 00:06:48.185 15:11:16 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@115 -- # spdk_tgt_pid=71204 00:06:48.185 15:11:16 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@116 -- # waitforlisten 71204 /var/tmp/spdk.sock 00:06:48.185 15:11:16 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@114 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:48.185 15:11:16 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 71204 ']' 00:06:48.185 15:11:16 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:48.185 15:11:16 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:48.185 15:11:16 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:48.185 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:48.185 15:11:16 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:48.185 15:11:16 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:48.185 [2024-11-27 15:11:16.183497] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:06:48.185 [2024-11-27 15:11:16.183622] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71204 ] 00:06:48.444 [2024-11-27 15:11:16.353719] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:48.444 [2024-11-27 15:11:16.379660] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:49.015 15:11:16 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:49.015 15:11:16 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 0 00:06:49.015 15:11:17 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@119 -- # spdk_tgt_pid2=71214 00:06:49.015 15:11:17 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@118 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:06:49.015 15:11:17 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@120 -- # NOT waitforlisten 71214 /var/tmp/spdk2.sock 00:06:49.015 15:11:17 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@652 -- # local es=0 00:06:49.015 15:11:17 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@654 -- # valid_exec_arg waitforlisten 71214 /var/tmp/spdk2.sock 00:06:49.015 15:11:17 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@640 -- # local arg=waitforlisten 00:06:49.015 15:11:17 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:49.015 15:11:17 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@644 -- # type -t waitforlisten 00:06:49.015 15:11:17 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:49.015 15:11:17 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@655 -- # waitforlisten 71214 /var/tmp/spdk2.sock 00:06:49.015 15:11:17 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 71214 ']' 00:06:49.015 15:11:17 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:49.015 15:11:17 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:49.015 15:11:17 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:49.015 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:49.015 15:11:17 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:49.015 15:11:17 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:49.015 [2024-11-27 15:11:17.094509] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:06:49.015 [2024-11-27 15:11:17.094704] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71214 ] 00:06:49.275 [2024-11-27 15:11:17.258963] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 0, probably process 71204 has claimed it. 00:06:49.275 [2024-11-27 15:11:17.259034] app.c: 912:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:06:49.843 ERROR: process (pid: 71214) is no longer running 00:06:49.843 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 850: kill: (71214) - No such process 00:06:49.843 15:11:17 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:49.843 15:11:17 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 1 00:06:49.843 15:11:17 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@655 -- # es=1 00:06:49.844 15:11:17 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:06:49.844 15:11:17 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:06:49.844 15:11:17 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:06:49.844 15:11:17 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@122 -- # locks_exist 71204 00:06:49.844 15:11:17 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 71204 00:06:49.844 15:11:17 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:06:50.104 15:11:18 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@124 -- # killprocess 71204 00:06:50.104 15:11:18 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # '[' -z 71204 ']' 00:06:50.104 15:11:18 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # kill -0 71204 00:06:50.104 15:11:18 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # uname 00:06:50.104 15:11:18 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:50.104 15:11:18 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71204 00:06:50.104 15:11:18 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:50.104 15:11:18 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:50.104 15:11:18 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71204' 00:06:50.104 killing process with pid 71204 00:06:50.104 15:11:18 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@973 -- # kill 71204 00:06:50.104 15:11:18 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@978 -- # wait 71204 00:06:50.363 00:06:50.363 real 0m2.351s 00:06:50.363 user 0m2.524s 00:06:50.363 sys 0m0.717s 00:06:50.363 15:11:18 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:50.363 15:11:18 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:50.363 ************************************ 00:06:50.363 END TEST locking_app_on_locked_coremask 00:06:50.363 ************************************ 00:06:50.622 15:11:18 event.cpu_locks -- event/cpu_locks.sh@171 -- # run_test locking_overlapped_coremask locking_overlapped_coremask 00:06:50.622 15:11:18 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:50.622 15:11:18 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:50.622 15:11:18 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:50.622 ************************************ 00:06:50.622 START TEST locking_overlapped_coremask 00:06:50.622 ************************************ 00:06:50.622 15:11:18 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1129 -- # locking_overlapped_coremask 00:06:50.622 15:11:18 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@132 -- # spdk_tgt_pid=71262 00:06:50.622 15:11:18 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@133 -- # waitforlisten 71262 /var/tmp/spdk.sock 00:06:50.622 15:11:18 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@131 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x7 00:06:50.622 15:11:18 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@835 -- # '[' -z 71262 ']' 00:06:50.622 15:11:18 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:50.622 15:11:18 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:50.622 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:50.622 15:11:18 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:50.622 15:11:18 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:50.622 15:11:18 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:50.622 [2024-11-27 15:11:18.607488] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:06:50.622 [2024-11-27 15:11:18.607620] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71262 ] 00:06:50.882 [2024-11-27 15:11:18.777576] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:06:50.882 [2024-11-27 15:11:18.805199] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:50.882 [2024-11-27 15:11:18.805278] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:50.882 [2024-11-27 15:11:18.805436] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:06:51.451 15:11:19 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:51.451 15:11:19 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@868 -- # return 0 00:06:51.451 15:11:19 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@136 -- # spdk_tgt_pid2=71280 00:06:51.451 15:11:19 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@135 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock 00:06:51.451 15:11:19 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@137 -- # NOT waitforlisten 71280 /var/tmp/spdk2.sock 00:06:51.451 15:11:19 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@652 -- # local es=0 00:06:51.451 15:11:19 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@654 -- # valid_exec_arg waitforlisten 71280 /var/tmp/spdk2.sock 00:06:51.451 15:11:19 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@640 -- # local arg=waitforlisten 00:06:51.451 15:11:19 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:51.451 15:11:19 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@644 -- # type -t waitforlisten 00:06:51.451 15:11:19 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:51.451 15:11:19 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@655 -- # waitforlisten 71280 /var/tmp/spdk2.sock 00:06:51.451 15:11:19 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@835 -- # '[' -z 71280 ']' 00:06:51.451 15:11:19 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:51.451 15:11:19 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:51.451 15:11:19 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:51.451 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:51.451 15:11:19 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:51.451 15:11:19 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:51.451 [2024-11-27 15:11:19.509846] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:06:51.451 [2024-11-27 15:11:19.510060] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71280 ] 00:06:51.711 [2024-11-27 15:11:19.673925] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 71262 has claimed it. 00:06:51.711 [2024-11-27 15:11:19.674013] app.c: 912:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:06:52.282 ERROR: process (pid: 71280) is no longer running 00:06:52.282 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 850: kill: (71280) - No such process 00:06:52.282 15:11:20 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:52.282 15:11:20 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@868 -- # return 1 00:06:52.282 15:11:20 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@655 -- # es=1 00:06:52.282 15:11:20 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:06:52.282 15:11:20 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:06:52.282 15:11:20 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:06:52.282 15:11:20 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@139 -- # check_remaining_locks 00:06:52.282 15:11:20 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:06:52.282 15:11:20 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:06:52.282 15:11:20 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:06:52.282 15:11:20 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@141 -- # killprocess 71262 00:06:52.282 15:11:20 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@954 -- # '[' -z 71262 ']' 00:06:52.282 15:11:20 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@958 -- # kill -0 71262 00:06:52.282 15:11:20 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@959 -- # uname 00:06:52.282 15:11:20 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:52.282 15:11:20 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71262 00:06:52.282 killing process with pid 71262 00:06:52.282 15:11:20 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:52.282 15:11:20 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:52.282 15:11:20 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71262' 00:06:52.282 15:11:20 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@973 -- # kill 71262 00:06:52.282 15:11:20 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@978 -- # wait 71262 00:06:52.543 00:06:52.543 real 0m2.035s 00:06:52.543 user 0m5.437s 00:06:52.543 sys 0m0.528s 00:06:52.543 ************************************ 00:06:52.543 END TEST locking_overlapped_coremask 00:06:52.543 ************************************ 00:06:52.543 15:11:20 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:52.543 15:11:20 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:52.543 15:11:20 event.cpu_locks -- event/cpu_locks.sh@172 -- # run_test locking_overlapped_coremask_via_rpc locking_overlapped_coremask_via_rpc 00:06:52.543 15:11:20 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:52.543 15:11:20 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:52.543 15:11:20 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:52.543 ************************************ 00:06:52.543 START TEST locking_overlapped_coremask_via_rpc 00:06:52.543 ************************************ 00:06:52.543 15:11:20 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1129 -- # locking_overlapped_coremask_via_rpc 00:06:52.543 15:11:20 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@148 -- # spdk_tgt_pid=71322 00:06:52.543 15:11:20 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@149 -- # waitforlisten 71322 /var/tmp/spdk.sock 00:06:52.543 15:11:20 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@147 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x7 --disable-cpumask-locks 00:06:52.543 15:11:20 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 71322 ']' 00:06:52.543 15:11:20 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:52.543 15:11:20 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:52.543 15:11:20 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:52.543 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:52.543 15:11:20 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:52.543 15:11:20 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:52.804 [2024-11-27 15:11:20.713112] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:06:52.804 [2024-11-27 15:11:20.713359] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71322 ] 00:06:52.804 [2024-11-27 15:11:20.885054] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:06:52.804 [2024-11-27 15:11:20.885192] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:06:53.063 [2024-11-27 15:11:20.913759] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:53.063 [2024-11-27 15:11:20.914022] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:53.063 [2024-11-27 15:11:20.914124] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:06:53.632 15:11:21 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:53.632 15:11:21 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:06:53.632 15:11:21 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@152 -- # spdk_tgt_pid2=71340 00:06:53.632 15:11:21 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@151 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock --disable-cpumask-locks 00:06:53.632 15:11:21 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@153 -- # waitforlisten 71340 /var/tmp/spdk2.sock 00:06:53.632 15:11:21 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 71340 ']' 00:06:53.632 15:11:21 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:53.632 15:11:21 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:53.632 15:11:21 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:53.632 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:53.632 15:11:21 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:53.632 15:11:21 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:53.632 [2024-11-27 15:11:21.570028] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:06:53.632 [2024-11-27 15:11:21.570160] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71340 ] 00:06:53.632 [2024-11-27 15:11:21.736135] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:06:53.632 [2024-11-27 15:11:21.736192] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:06:53.891 [2024-11-27 15:11:21.798238] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:06:53.891 [2024-11-27 15:11:21.798313] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:06:53.891 [2024-11-27 15:11:21.798436] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:06:54.459 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:54.459 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:06:54.459 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@155 -- # rpc_cmd framework_enable_cpumask_locks 00:06:54.459 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:54.459 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:54.459 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:54.459 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@156 -- # NOT rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:06:54.459 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@652 -- # local es=0 00:06:54.459 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:06:54.459 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:06:54.459 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:54.459 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:06:54.459 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:54.459 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@655 -- # rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:06:54.459 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:54.459 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:54.459 [2024-11-27 15:11:22.464082] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 71322 has claimed it. 00:06:54.459 request: 00:06:54.459 { 00:06:54.459 "method": "framework_enable_cpumask_locks", 00:06:54.459 "req_id": 1 00:06:54.459 } 00:06:54.459 Got JSON-RPC error response 00:06:54.459 response: 00:06:54.459 { 00:06:54.459 "code": -32603, 00:06:54.459 "message": "Failed to claim CPU core: 2" 00:06:54.459 } 00:06:54.459 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:06:54.459 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@655 -- # es=1 00:06:54.459 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:06:54.459 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:06:54.459 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:06:54.459 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@158 -- # waitforlisten 71322 /var/tmp/spdk.sock 00:06:54.459 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 71322 ']' 00:06:54.459 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:54.459 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:54.459 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:54.459 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:54.459 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:54.459 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:54.718 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:54.718 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:06:54.718 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@159 -- # waitforlisten 71340 /var/tmp/spdk2.sock 00:06:54.718 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 71340 ']' 00:06:54.718 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:54.718 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:54.718 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:54.718 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:54.718 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:54.718 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:54.978 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:54.978 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:06:54.978 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@161 -- # check_remaining_locks 00:06:54.978 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:06:54.978 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:06:54.978 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:06:54.978 00:06:54.978 real 0m2.285s 00:06:54.978 user 0m1.068s 00:06:54.978 sys 0m0.149s 00:06:54.978 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:54.978 15:11:22 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:54.978 ************************************ 00:06:54.978 END TEST locking_overlapped_coremask_via_rpc 00:06:54.978 ************************************ 00:06:54.978 15:11:22 event.cpu_locks -- event/cpu_locks.sh@174 -- # cleanup 00:06:54.978 15:11:22 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 71322 ]] 00:06:54.978 15:11:22 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 71322 00:06:54.978 15:11:22 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 71322 ']' 00:06:54.978 15:11:22 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 71322 00:06:54.978 15:11:22 event.cpu_locks -- common/autotest_common.sh@959 -- # uname 00:06:54.978 15:11:22 event.cpu_locks -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:54.978 15:11:22 event.cpu_locks -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71322 00:06:54.978 15:11:22 event.cpu_locks -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:54.978 killing process with pid 71322 00:06:54.978 15:11:22 event.cpu_locks -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:54.978 15:11:22 event.cpu_locks -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71322' 00:06:54.978 15:11:22 event.cpu_locks -- common/autotest_common.sh@973 -- # kill 71322 00:06:54.978 15:11:22 event.cpu_locks -- common/autotest_common.sh@978 -- # wait 71322 00:06:55.547 15:11:23 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 71340 ]] 00:06:55.547 15:11:23 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 71340 00:06:55.547 15:11:23 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 71340 ']' 00:06:55.547 15:11:23 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 71340 00:06:55.547 15:11:23 event.cpu_locks -- common/autotest_common.sh@959 -- # uname 00:06:55.547 15:11:23 event.cpu_locks -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:55.547 15:11:23 event.cpu_locks -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71340 00:06:55.547 killing process with pid 71340 00:06:55.547 15:11:23 event.cpu_locks -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:06:55.547 15:11:23 event.cpu_locks -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:06:55.547 15:11:23 event.cpu_locks -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71340' 00:06:55.547 15:11:23 event.cpu_locks -- common/autotest_common.sh@973 -- # kill 71340 00:06:55.547 15:11:23 event.cpu_locks -- common/autotest_common.sh@978 -- # wait 71340 00:06:55.807 15:11:23 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:06:55.807 Process with pid 71322 is not found 00:06:55.807 15:11:23 event.cpu_locks -- event/cpu_locks.sh@1 -- # cleanup 00:06:55.807 15:11:23 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 71322 ]] 00:06:55.807 15:11:23 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 71322 00:06:55.807 15:11:23 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 71322 ']' 00:06:55.807 15:11:23 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 71322 00:06:55.807 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 958: kill: (71322) - No such process 00:06:55.807 15:11:23 event.cpu_locks -- common/autotest_common.sh@981 -- # echo 'Process with pid 71322 is not found' 00:06:55.807 15:11:23 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 71340 ]] 00:06:55.807 15:11:23 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 71340 00:06:55.807 15:11:23 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 71340 ']' 00:06:55.808 15:11:23 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 71340 00:06:55.808 Process with pid 71340 is not found 00:06:55.808 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 958: kill: (71340) - No such process 00:06:55.808 15:11:23 event.cpu_locks -- common/autotest_common.sh@981 -- # echo 'Process with pid 71340 is not found' 00:06:55.808 15:11:23 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:06:55.808 ************************************ 00:06:55.808 END TEST cpu_locks 00:06:55.808 ************************************ 00:06:55.808 00:06:55.808 real 0m18.689s 00:06:55.808 user 0m31.106s 00:06:55.808 sys 0m5.894s 00:06:55.808 15:11:23 event.cpu_locks -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:55.808 15:11:23 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:55.808 ************************************ 00:06:55.808 END TEST event 00:06:55.808 ************************************ 00:06:55.808 00:06:55.808 real 0m45.135s 00:06:55.808 user 1m24.301s 00:06:55.808 sys 0m9.397s 00:06:55.808 15:11:23 event -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:55.808 15:11:23 event -- common/autotest_common.sh@10 -- # set +x 00:06:55.808 15:11:23 -- spdk/autotest.sh@169 -- # run_test thread /home/vagrant/spdk_repo/spdk/test/thread/thread.sh 00:06:55.808 15:11:23 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:55.808 15:11:23 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:55.808 15:11:23 -- common/autotest_common.sh@10 -- # set +x 00:06:55.808 ************************************ 00:06:55.808 START TEST thread 00:06:55.808 ************************************ 00:06:55.808 15:11:23 thread -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/thread/thread.sh 00:06:56.067 * Looking for test storage... 00:06:56.067 * Found test storage at /home/vagrant/spdk_repo/spdk/test/thread 00:06:56.067 15:11:24 thread -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:06:56.067 15:11:24 thread -- common/autotest_common.sh@1693 -- # lcov --version 00:06:56.067 15:11:24 thread -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:06:56.067 15:11:24 thread -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:06:56.067 15:11:24 thread -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:56.067 15:11:24 thread -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:56.067 15:11:24 thread -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:56.067 15:11:24 thread -- scripts/common.sh@336 -- # IFS=.-: 00:06:56.067 15:11:24 thread -- scripts/common.sh@336 -- # read -ra ver1 00:06:56.067 15:11:24 thread -- scripts/common.sh@337 -- # IFS=.-: 00:06:56.067 15:11:24 thread -- scripts/common.sh@337 -- # read -ra ver2 00:06:56.067 15:11:24 thread -- scripts/common.sh@338 -- # local 'op=<' 00:06:56.067 15:11:24 thread -- scripts/common.sh@340 -- # ver1_l=2 00:06:56.067 15:11:24 thread -- scripts/common.sh@341 -- # ver2_l=1 00:06:56.067 15:11:24 thread -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:56.067 15:11:24 thread -- scripts/common.sh@344 -- # case "$op" in 00:06:56.067 15:11:24 thread -- scripts/common.sh@345 -- # : 1 00:06:56.067 15:11:24 thread -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:56.067 15:11:24 thread -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:56.067 15:11:24 thread -- scripts/common.sh@365 -- # decimal 1 00:06:56.067 15:11:24 thread -- scripts/common.sh@353 -- # local d=1 00:06:56.067 15:11:24 thread -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:56.067 15:11:24 thread -- scripts/common.sh@355 -- # echo 1 00:06:56.067 15:11:24 thread -- scripts/common.sh@365 -- # ver1[v]=1 00:06:56.067 15:11:24 thread -- scripts/common.sh@366 -- # decimal 2 00:06:56.067 15:11:24 thread -- scripts/common.sh@353 -- # local d=2 00:06:56.067 15:11:24 thread -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:56.067 15:11:24 thread -- scripts/common.sh@355 -- # echo 2 00:06:56.067 15:11:24 thread -- scripts/common.sh@366 -- # ver2[v]=2 00:06:56.067 15:11:24 thread -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:56.067 15:11:24 thread -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:56.067 15:11:24 thread -- scripts/common.sh@368 -- # return 0 00:06:56.067 15:11:24 thread -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:56.067 15:11:24 thread -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:06:56.067 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:56.067 --rc genhtml_branch_coverage=1 00:06:56.067 --rc genhtml_function_coverage=1 00:06:56.067 --rc genhtml_legend=1 00:06:56.067 --rc geninfo_all_blocks=1 00:06:56.067 --rc geninfo_unexecuted_blocks=1 00:06:56.067 00:06:56.067 ' 00:06:56.067 15:11:24 thread -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:06:56.067 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:56.067 --rc genhtml_branch_coverage=1 00:06:56.067 --rc genhtml_function_coverage=1 00:06:56.067 --rc genhtml_legend=1 00:06:56.067 --rc geninfo_all_blocks=1 00:06:56.067 --rc geninfo_unexecuted_blocks=1 00:06:56.067 00:06:56.067 ' 00:06:56.067 15:11:24 thread -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:06:56.067 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:56.067 --rc genhtml_branch_coverage=1 00:06:56.067 --rc genhtml_function_coverage=1 00:06:56.067 --rc genhtml_legend=1 00:06:56.067 --rc geninfo_all_blocks=1 00:06:56.067 --rc geninfo_unexecuted_blocks=1 00:06:56.067 00:06:56.067 ' 00:06:56.067 15:11:24 thread -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:06:56.067 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:56.067 --rc genhtml_branch_coverage=1 00:06:56.067 --rc genhtml_function_coverage=1 00:06:56.067 --rc genhtml_legend=1 00:06:56.067 --rc geninfo_all_blocks=1 00:06:56.067 --rc geninfo_unexecuted_blocks=1 00:06:56.067 00:06:56.067 ' 00:06:56.067 15:11:24 thread -- thread/thread.sh@11 -- # run_test thread_poller_perf /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:06:56.067 15:11:24 thread -- common/autotest_common.sh@1105 -- # '[' 8 -le 1 ']' 00:06:56.067 15:11:24 thread -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:56.067 15:11:24 thread -- common/autotest_common.sh@10 -- # set +x 00:06:56.067 ************************************ 00:06:56.067 START TEST thread_poller_perf 00:06:56.067 ************************************ 00:06:56.067 15:11:24 thread.thread_poller_perf -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:06:56.326 [2024-11-27 15:11:24.172076] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:06:56.326 [2024-11-27 15:11:24.172193] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71469 ] 00:06:56.326 [2024-11-27 15:11:24.344069] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:56.326 [2024-11-27 15:11:24.369710] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:56.326 Running 1000 pollers for 1 seconds with 1 microseconds period. 00:06:57.706 [2024-11-27T15:11:25.813Z] ====================================== 00:06:57.706 [2024-11-27T15:11:25.813Z] busy:2297497108 (cyc) 00:06:57.706 [2024-11-27T15:11:25.813Z] total_run_count: 421000 00:06:57.706 [2024-11-27T15:11:25.813Z] tsc_hz: 2290000000 (cyc) 00:06:57.706 [2024-11-27T15:11:25.813Z] ====================================== 00:06:57.706 [2024-11-27T15:11:25.813Z] poller_cost: 5457 (cyc), 2382 (nsec) 00:06:57.706 ************************************ 00:06:57.706 END TEST thread_poller_perf 00:06:57.706 ************************************ 00:06:57.706 00:06:57.706 real 0m1.310s 00:06:57.706 user 0m1.123s 00:06:57.706 sys 0m0.081s 00:06:57.706 15:11:25 thread.thread_poller_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:57.706 15:11:25 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:06:57.706 15:11:25 thread -- thread/thread.sh@12 -- # run_test thread_poller_perf /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:06:57.706 15:11:25 thread -- common/autotest_common.sh@1105 -- # '[' 8 -le 1 ']' 00:06:57.706 15:11:25 thread -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:57.706 15:11:25 thread -- common/autotest_common.sh@10 -- # set +x 00:06:57.706 ************************************ 00:06:57.706 START TEST thread_poller_perf 00:06:57.706 ************************************ 00:06:57.706 15:11:25 thread.thread_poller_perf -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:06:57.706 [2024-11-27 15:11:25.541775] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:06:57.706 [2024-11-27 15:11:25.541902] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71511 ] 00:06:57.706 [2024-11-27 15:11:25.710999] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:57.706 Running 1000 pollers for 1 seconds with 0 microseconds period. 00:06:57.706 [2024-11-27 15:11:25.736615] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:59.087 [2024-11-27T15:11:27.194Z] ====================================== 00:06:59.087 [2024-11-27T15:11:27.194Z] busy:2293619134 (cyc) 00:06:59.087 [2024-11-27T15:11:27.194Z] total_run_count: 5303000 00:06:59.087 [2024-11-27T15:11:27.194Z] tsc_hz: 2290000000 (cyc) 00:06:59.087 [2024-11-27T15:11:27.194Z] ====================================== 00:06:59.087 [2024-11-27T15:11:27.194Z] poller_cost: 432 (cyc), 188 (nsec) 00:06:59.087 00:06:59.087 real 0m1.302s 00:06:59.087 user 0m1.107s 00:06:59.087 sys 0m0.089s 00:06:59.087 ************************************ 00:06:59.087 END TEST thread_poller_perf 00:06:59.087 ************************************ 00:06:59.087 15:11:26 thread.thread_poller_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:59.087 15:11:26 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:06:59.087 15:11:26 thread -- thread/thread.sh@17 -- # [[ y != \y ]] 00:06:59.087 00:06:59.087 real 0m2.955s 00:06:59.087 user 0m2.398s 00:06:59.087 sys 0m0.358s 00:06:59.087 ************************************ 00:06:59.087 END TEST thread 00:06:59.087 ************************************ 00:06:59.087 15:11:26 thread -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:59.087 15:11:26 thread -- common/autotest_common.sh@10 -- # set +x 00:06:59.087 15:11:26 -- spdk/autotest.sh@171 -- # [[ 0 -eq 1 ]] 00:06:59.087 15:11:26 -- spdk/autotest.sh@176 -- # run_test app_cmdline /home/vagrant/spdk_repo/spdk/test/app/cmdline.sh 00:06:59.087 15:11:26 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:59.087 15:11:26 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:59.087 15:11:26 -- common/autotest_common.sh@10 -- # set +x 00:06:59.087 ************************************ 00:06:59.087 START TEST app_cmdline 00:06:59.087 ************************************ 00:06:59.087 15:11:26 app_cmdline -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/app/cmdline.sh 00:06:59.087 * Looking for test storage... 00:06:59.087 * Found test storage at /home/vagrant/spdk_repo/spdk/test/app 00:06:59.087 15:11:27 app_cmdline -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:06:59.087 15:11:27 app_cmdline -- common/autotest_common.sh@1693 -- # lcov --version 00:06:59.087 15:11:27 app_cmdline -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:06:59.087 15:11:27 app_cmdline -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:06:59.087 15:11:27 app_cmdline -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:59.087 15:11:27 app_cmdline -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:59.087 15:11:27 app_cmdline -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:59.087 15:11:27 app_cmdline -- scripts/common.sh@336 -- # IFS=.-: 00:06:59.087 15:11:27 app_cmdline -- scripts/common.sh@336 -- # read -ra ver1 00:06:59.087 15:11:27 app_cmdline -- scripts/common.sh@337 -- # IFS=.-: 00:06:59.087 15:11:27 app_cmdline -- scripts/common.sh@337 -- # read -ra ver2 00:06:59.087 15:11:27 app_cmdline -- scripts/common.sh@338 -- # local 'op=<' 00:06:59.087 15:11:27 app_cmdline -- scripts/common.sh@340 -- # ver1_l=2 00:06:59.087 15:11:27 app_cmdline -- scripts/common.sh@341 -- # ver2_l=1 00:06:59.087 15:11:27 app_cmdline -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:59.088 15:11:27 app_cmdline -- scripts/common.sh@344 -- # case "$op" in 00:06:59.088 15:11:27 app_cmdline -- scripts/common.sh@345 -- # : 1 00:06:59.088 15:11:27 app_cmdline -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:59.088 15:11:27 app_cmdline -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:59.088 15:11:27 app_cmdline -- scripts/common.sh@365 -- # decimal 1 00:06:59.088 15:11:27 app_cmdline -- scripts/common.sh@353 -- # local d=1 00:06:59.088 15:11:27 app_cmdline -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:59.088 15:11:27 app_cmdline -- scripts/common.sh@355 -- # echo 1 00:06:59.088 15:11:27 app_cmdline -- scripts/common.sh@365 -- # ver1[v]=1 00:06:59.088 15:11:27 app_cmdline -- scripts/common.sh@366 -- # decimal 2 00:06:59.088 15:11:27 app_cmdline -- scripts/common.sh@353 -- # local d=2 00:06:59.088 15:11:27 app_cmdline -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:59.088 15:11:27 app_cmdline -- scripts/common.sh@355 -- # echo 2 00:06:59.088 15:11:27 app_cmdline -- scripts/common.sh@366 -- # ver2[v]=2 00:06:59.088 15:11:27 app_cmdline -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:59.088 15:11:27 app_cmdline -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:59.088 15:11:27 app_cmdline -- scripts/common.sh@368 -- # return 0 00:06:59.088 15:11:27 app_cmdline -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:59.088 15:11:27 app_cmdline -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:06:59.088 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:59.088 --rc genhtml_branch_coverage=1 00:06:59.088 --rc genhtml_function_coverage=1 00:06:59.088 --rc genhtml_legend=1 00:06:59.088 --rc geninfo_all_blocks=1 00:06:59.088 --rc geninfo_unexecuted_blocks=1 00:06:59.088 00:06:59.088 ' 00:06:59.088 15:11:27 app_cmdline -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:06:59.088 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:59.088 --rc genhtml_branch_coverage=1 00:06:59.088 --rc genhtml_function_coverage=1 00:06:59.088 --rc genhtml_legend=1 00:06:59.088 --rc geninfo_all_blocks=1 00:06:59.088 --rc geninfo_unexecuted_blocks=1 00:06:59.088 00:06:59.088 ' 00:06:59.088 15:11:27 app_cmdline -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:06:59.088 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:59.088 --rc genhtml_branch_coverage=1 00:06:59.088 --rc genhtml_function_coverage=1 00:06:59.088 --rc genhtml_legend=1 00:06:59.088 --rc geninfo_all_blocks=1 00:06:59.088 --rc geninfo_unexecuted_blocks=1 00:06:59.088 00:06:59.088 ' 00:06:59.088 15:11:27 app_cmdline -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:06:59.088 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:59.088 --rc genhtml_branch_coverage=1 00:06:59.088 --rc genhtml_function_coverage=1 00:06:59.088 --rc genhtml_legend=1 00:06:59.088 --rc geninfo_all_blocks=1 00:06:59.088 --rc geninfo_unexecuted_blocks=1 00:06:59.088 00:06:59.088 ' 00:06:59.088 15:11:27 app_cmdline -- app/cmdline.sh@14 -- # trap 'killprocess $spdk_tgt_pid' EXIT 00:06:59.088 15:11:27 app_cmdline -- app/cmdline.sh@17 -- # spdk_tgt_pid=71589 00:06:59.088 15:11:27 app_cmdline -- app/cmdline.sh@18 -- # waitforlisten 71589 00:06:59.088 15:11:27 app_cmdline -- app/cmdline.sh@16 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --rpcs-allowed spdk_get_version,rpc_get_methods 00:06:59.088 15:11:27 app_cmdline -- common/autotest_common.sh@835 -- # '[' -z 71589 ']' 00:06:59.088 15:11:27 app_cmdline -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:59.088 15:11:27 app_cmdline -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:59.088 15:11:27 app_cmdline -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:59.088 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:59.088 15:11:27 app_cmdline -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:59.088 15:11:27 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:06:59.348 [2024-11-27 15:11:27.235399] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:06:59.348 [2024-11-27 15:11:27.235613] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71589 ] 00:06:59.348 [2024-11-27 15:11:27.406558] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:59.348 [2024-11-27 15:11:27.431428] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:00.286 15:11:28 app_cmdline -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:00.286 15:11:28 app_cmdline -- common/autotest_common.sh@868 -- # return 0 00:07:00.286 15:11:28 app_cmdline -- app/cmdline.sh@20 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py spdk_get_version 00:07:00.286 { 00:07:00.286 "version": "SPDK v25.01-pre git sha1 35cd3e84d", 00:07:00.286 "fields": { 00:07:00.286 "major": 25, 00:07:00.286 "minor": 1, 00:07:00.286 "patch": 0, 00:07:00.286 "suffix": "-pre", 00:07:00.286 "commit": "35cd3e84d" 00:07:00.286 } 00:07:00.286 } 00:07:00.286 15:11:28 app_cmdline -- app/cmdline.sh@22 -- # expected_methods=() 00:07:00.286 15:11:28 app_cmdline -- app/cmdline.sh@23 -- # expected_methods+=("rpc_get_methods") 00:07:00.287 15:11:28 app_cmdline -- app/cmdline.sh@24 -- # expected_methods+=("spdk_get_version") 00:07:00.287 15:11:28 app_cmdline -- app/cmdline.sh@26 -- # methods=($(rpc_cmd rpc_get_methods | jq -r ".[]" | sort)) 00:07:00.287 15:11:28 app_cmdline -- app/cmdline.sh@26 -- # rpc_cmd rpc_get_methods 00:07:00.287 15:11:28 app_cmdline -- app/cmdline.sh@26 -- # jq -r '.[]' 00:07:00.287 15:11:28 app_cmdline -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:00.287 15:11:28 app_cmdline -- app/cmdline.sh@26 -- # sort 00:07:00.287 15:11:28 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:07:00.287 15:11:28 app_cmdline -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:00.287 15:11:28 app_cmdline -- app/cmdline.sh@27 -- # (( 2 == 2 )) 00:07:00.287 15:11:28 app_cmdline -- app/cmdline.sh@28 -- # [[ rpc_get_methods spdk_get_version == \r\p\c\_\g\e\t\_\m\e\t\h\o\d\s\ \s\p\d\k\_\g\e\t\_\v\e\r\s\i\o\n ]] 00:07:00.287 15:11:28 app_cmdline -- app/cmdline.sh@30 -- # NOT /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:07:00.287 15:11:28 app_cmdline -- common/autotest_common.sh@652 -- # local es=0 00:07:00.287 15:11:28 app_cmdline -- common/autotest_common.sh@654 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:07:00.287 15:11:28 app_cmdline -- common/autotest_common.sh@640 -- # local arg=/home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:07:00.287 15:11:28 app_cmdline -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:00.287 15:11:28 app_cmdline -- common/autotest_common.sh@644 -- # type -t /home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:07:00.287 15:11:28 app_cmdline -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:00.287 15:11:28 app_cmdline -- common/autotest_common.sh@646 -- # type -P /home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:07:00.287 15:11:28 app_cmdline -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:00.287 15:11:28 app_cmdline -- common/autotest_common.sh@646 -- # arg=/home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:07:00.287 15:11:28 app_cmdline -- common/autotest_common.sh@646 -- # [[ -x /home/vagrant/spdk_repo/spdk/scripts/rpc.py ]] 00:07:00.287 15:11:28 app_cmdline -- common/autotest_common.sh@655 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:07:00.545 request: 00:07:00.545 { 00:07:00.545 "method": "env_dpdk_get_mem_stats", 00:07:00.545 "req_id": 1 00:07:00.545 } 00:07:00.545 Got JSON-RPC error response 00:07:00.545 response: 00:07:00.545 { 00:07:00.545 "code": -32601, 00:07:00.545 "message": "Method not found" 00:07:00.545 } 00:07:00.545 15:11:28 app_cmdline -- common/autotest_common.sh@655 -- # es=1 00:07:00.545 15:11:28 app_cmdline -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:07:00.545 15:11:28 app_cmdline -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:07:00.545 15:11:28 app_cmdline -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:07:00.545 15:11:28 app_cmdline -- app/cmdline.sh@1 -- # killprocess 71589 00:07:00.545 15:11:28 app_cmdline -- common/autotest_common.sh@954 -- # '[' -z 71589 ']' 00:07:00.546 15:11:28 app_cmdline -- common/autotest_common.sh@958 -- # kill -0 71589 00:07:00.546 15:11:28 app_cmdline -- common/autotest_common.sh@959 -- # uname 00:07:00.546 15:11:28 app_cmdline -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:00.546 15:11:28 app_cmdline -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71589 00:07:00.546 killing process with pid 71589 00:07:00.546 15:11:28 app_cmdline -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:00.546 15:11:28 app_cmdline -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:00.546 15:11:28 app_cmdline -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71589' 00:07:00.546 15:11:28 app_cmdline -- common/autotest_common.sh@973 -- # kill 71589 00:07:00.546 15:11:28 app_cmdline -- common/autotest_common.sh@978 -- # wait 71589 00:07:00.805 ************************************ 00:07:00.805 END TEST app_cmdline 00:07:00.805 ************************************ 00:07:00.806 00:07:00.806 real 0m1.941s 00:07:00.806 user 0m2.153s 00:07:00.806 sys 0m0.531s 00:07:00.806 15:11:28 app_cmdline -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:00.806 15:11:28 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:07:01.066 15:11:28 -- spdk/autotest.sh@177 -- # run_test version /home/vagrant/spdk_repo/spdk/test/app/version.sh 00:07:01.066 15:11:28 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:01.066 15:11:28 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:01.066 15:11:28 -- common/autotest_common.sh@10 -- # set +x 00:07:01.066 ************************************ 00:07:01.066 START TEST version 00:07:01.066 ************************************ 00:07:01.066 15:11:28 version -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/app/version.sh 00:07:01.066 * Looking for test storage... 00:07:01.066 * Found test storage at /home/vagrant/spdk_repo/spdk/test/app 00:07:01.066 15:11:29 version -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:07:01.066 15:11:29 version -- common/autotest_common.sh@1693 -- # lcov --version 00:07:01.066 15:11:29 version -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:07:01.066 15:11:29 version -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:07:01.066 15:11:29 version -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:01.066 15:11:29 version -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:01.066 15:11:29 version -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:01.066 15:11:29 version -- scripts/common.sh@336 -- # IFS=.-: 00:07:01.066 15:11:29 version -- scripts/common.sh@336 -- # read -ra ver1 00:07:01.066 15:11:29 version -- scripts/common.sh@337 -- # IFS=.-: 00:07:01.066 15:11:29 version -- scripts/common.sh@337 -- # read -ra ver2 00:07:01.066 15:11:29 version -- scripts/common.sh@338 -- # local 'op=<' 00:07:01.066 15:11:29 version -- scripts/common.sh@340 -- # ver1_l=2 00:07:01.066 15:11:29 version -- scripts/common.sh@341 -- # ver2_l=1 00:07:01.066 15:11:29 version -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:01.066 15:11:29 version -- scripts/common.sh@344 -- # case "$op" in 00:07:01.066 15:11:29 version -- scripts/common.sh@345 -- # : 1 00:07:01.066 15:11:29 version -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:01.066 15:11:29 version -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:01.066 15:11:29 version -- scripts/common.sh@365 -- # decimal 1 00:07:01.066 15:11:29 version -- scripts/common.sh@353 -- # local d=1 00:07:01.066 15:11:29 version -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:01.066 15:11:29 version -- scripts/common.sh@355 -- # echo 1 00:07:01.066 15:11:29 version -- scripts/common.sh@365 -- # ver1[v]=1 00:07:01.066 15:11:29 version -- scripts/common.sh@366 -- # decimal 2 00:07:01.066 15:11:29 version -- scripts/common.sh@353 -- # local d=2 00:07:01.066 15:11:29 version -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:01.066 15:11:29 version -- scripts/common.sh@355 -- # echo 2 00:07:01.066 15:11:29 version -- scripts/common.sh@366 -- # ver2[v]=2 00:07:01.066 15:11:29 version -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:01.066 15:11:29 version -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:01.066 15:11:29 version -- scripts/common.sh@368 -- # return 0 00:07:01.066 15:11:29 version -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:01.066 15:11:29 version -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:07:01.066 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:01.066 --rc genhtml_branch_coverage=1 00:07:01.066 --rc genhtml_function_coverage=1 00:07:01.066 --rc genhtml_legend=1 00:07:01.066 --rc geninfo_all_blocks=1 00:07:01.066 --rc geninfo_unexecuted_blocks=1 00:07:01.066 00:07:01.066 ' 00:07:01.066 15:11:29 version -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:07:01.066 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:01.066 --rc genhtml_branch_coverage=1 00:07:01.066 --rc genhtml_function_coverage=1 00:07:01.066 --rc genhtml_legend=1 00:07:01.066 --rc geninfo_all_blocks=1 00:07:01.066 --rc geninfo_unexecuted_blocks=1 00:07:01.066 00:07:01.066 ' 00:07:01.066 15:11:29 version -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:07:01.066 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:01.066 --rc genhtml_branch_coverage=1 00:07:01.066 --rc genhtml_function_coverage=1 00:07:01.066 --rc genhtml_legend=1 00:07:01.066 --rc geninfo_all_blocks=1 00:07:01.066 --rc geninfo_unexecuted_blocks=1 00:07:01.066 00:07:01.066 ' 00:07:01.066 15:11:29 version -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:07:01.066 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:01.066 --rc genhtml_branch_coverage=1 00:07:01.066 --rc genhtml_function_coverage=1 00:07:01.066 --rc genhtml_legend=1 00:07:01.066 --rc geninfo_all_blocks=1 00:07:01.066 --rc geninfo_unexecuted_blocks=1 00:07:01.066 00:07:01.066 ' 00:07:01.066 15:11:29 version -- app/version.sh@17 -- # get_header_version major 00:07:01.066 15:11:29 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MAJOR[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:07:01.066 15:11:29 version -- app/version.sh@14 -- # cut -f2 00:07:01.066 15:11:29 version -- app/version.sh@14 -- # tr -d '"' 00:07:01.066 15:11:29 version -- app/version.sh@17 -- # major=25 00:07:01.066 15:11:29 version -- app/version.sh@18 -- # get_header_version minor 00:07:01.066 15:11:29 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MINOR[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:07:01.066 15:11:29 version -- app/version.sh@14 -- # cut -f2 00:07:01.066 15:11:29 version -- app/version.sh@14 -- # tr -d '"' 00:07:01.325 15:11:29 version -- app/version.sh@18 -- # minor=1 00:07:01.325 15:11:29 version -- app/version.sh@19 -- # get_header_version patch 00:07:01.325 15:11:29 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_PATCH[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:07:01.325 15:11:29 version -- app/version.sh@14 -- # cut -f2 00:07:01.325 15:11:29 version -- app/version.sh@14 -- # tr -d '"' 00:07:01.325 15:11:29 version -- app/version.sh@19 -- # patch=0 00:07:01.325 15:11:29 version -- app/version.sh@20 -- # get_header_version suffix 00:07:01.325 15:11:29 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_SUFFIX[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:07:01.325 15:11:29 version -- app/version.sh@14 -- # cut -f2 00:07:01.325 15:11:29 version -- app/version.sh@14 -- # tr -d '"' 00:07:01.325 15:11:29 version -- app/version.sh@20 -- # suffix=-pre 00:07:01.325 15:11:29 version -- app/version.sh@22 -- # version=25.1 00:07:01.325 15:11:29 version -- app/version.sh@25 -- # (( patch != 0 )) 00:07:01.325 15:11:29 version -- app/version.sh@28 -- # version=25.1rc0 00:07:01.325 15:11:29 version -- app/version.sh@30 -- # PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python 00:07:01.325 15:11:29 version -- app/version.sh@30 -- # python3 -c 'import spdk; print(spdk.__version__)' 00:07:01.325 15:11:29 version -- app/version.sh@30 -- # py_version=25.1rc0 00:07:01.325 15:11:29 version -- app/version.sh@31 -- # [[ 25.1rc0 == \2\5\.\1\r\c\0 ]] 00:07:01.325 00:07:01.325 real 0m0.315s 00:07:01.325 user 0m0.184s 00:07:01.325 sys 0m0.188s 00:07:01.325 ************************************ 00:07:01.325 END TEST version 00:07:01.325 ************************************ 00:07:01.325 15:11:29 version -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:01.325 15:11:29 version -- common/autotest_common.sh@10 -- # set +x 00:07:01.325 15:11:29 -- spdk/autotest.sh@179 -- # '[' 0 -eq 1 ']' 00:07:01.325 15:11:29 -- spdk/autotest.sh@188 -- # [[ 1 -eq 1 ]] 00:07:01.325 15:11:29 -- spdk/autotest.sh@189 -- # run_test bdev_raid /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh 00:07:01.325 15:11:29 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:01.325 15:11:29 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:01.325 15:11:29 -- common/autotest_common.sh@10 -- # set +x 00:07:01.325 ************************************ 00:07:01.325 START TEST bdev_raid 00:07:01.325 ************************************ 00:07:01.325 15:11:29 bdev_raid -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh 00:07:01.325 * Looking for test storage... 00:07:01.585 * Found test storage at /home/vagrant/spdk_repo/spdk/test/bdev 00:07:01.585 15:11:29 bdev_raid -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:07:01.585 15:11:29 bdev_raid -- common/autotest_common.sh@1693 -- # lcov --version 00:07:01.585 15:11:29 bdev_raid -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:07:01.585 15:11:29 bdev_raid -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:07:01.585 15:11:29 bdev_raid -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:01.585 15:11:29 bdev_raid -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:01.585 15:11:29 bdev_raid -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:01.585 15:11:29 bdev_raid -- scripts/common.sh@336 -- # IFS=.-: 00:07:01.585 15:11:29 bdev_raid -- scripts/common.sh@336 -- # read -ra ver1 00:07:01.585 15:11:29 bdev_raid -- scripts/common.sh@337 -- # IFS=.-: 00:07:01.585 15:11:29 bdev_raid -- scripts/common.sh@337 -- # read -ra ver2 00:07:01.585 15:11:29 bdev_raid -- scripts/common.sh@338 -- # local 'op=<' 00:07:01.585 15:11:29 bdev_raid -- scripts/common.sh@340 -- # ver1_l=2 00:07:01.585 15:11:29 bdev_raid -- scripts/common.sh@341 -- # ver2_l=1 00:07:01.585 15:11:29 bdev_raid -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:01.585 15:11:29 bdev_raid -- scripts/common.sh@344 -- # case "$op" in 00:07:01.585 15:11:29 bdev_raid -- scripts/common.sh@345 -- # : 1 00:07:01.585 15:11:29 bdev_raid -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:01.585 15:11:29 bdev_raid -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:01.586 15:11:29 bdev_raid -- scripts/common.sh@365 -- # decimal 1 00:07:01.586 15:11:29 bdev_raid -- scripts/common.sh@353 -- # local d=1 00:07:01.586 15:11:29 bdev_raid -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:01.586 15:11:29 bdev_raid -- scripts/common.sh@355 -- # echo 1 00:07:01.586 15:11:29 bdev_raid -- scripts/common.sh@365 -- # ver1[v]=1 00:07:01.586 15:11:29 bdev_raid -- scripts/common.sh@366 -- # decimal 2 00:07:01.586 15:11:29 bdev_raid -- scripts/common.sh@353 -- # local d=2 00:07:01.586 15:11:29 bdev_raid -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:01.586 15:11:29 bdev_raid -- scripts/common.sh@355 -- # echo 2 00:07:01.586 15:11:29 bdev_raid -- scripts/common.sh@366 -- # ver2[v]=2 00:07:01.586 15:11:29 bdev_raid -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:01.586 15:11:29 bdev_raid -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:01.586 15:11:29 bdev_raid -- scripts/common.sh@368 -- # return 0 00:07:01.586 15:11:29 bdev_raid -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:01.586 15:11:29 bdev_raid -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:07:01.586 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:01.586 --rc genhtml_branch_coverage=1 00:07:01.586 --rc genhtml_function_coverage=1 00:07:01.586 --rc genhtml_legend=1 00:07:01.586 --rc geninfo_all_blocks=1 00:07:01.586 --rc geninfo_unexecuted_blocks=1 00:07:01.586 00:07:01.586 ' 00:07:01.586 15:11:29 bdev_raid -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:07:01.586 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:01.586 --rc genhtml_branch_coverage=1 00:07:01.586 --rc genhtml_function_coverage=1 00:07:01.586 --rc genhtml_legend=1 00:07:01.586 --rc geninfo_all_blocks=1 00:07:01.586 --rc geninfo_unexecuted_blocks=1 00:07:01.586 00:07:01.586 ' 00:07:01.586 15:11:29 bdev_raid -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:07:01.586 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:01.586 --rc genhtml_branch_coverage=1 00:07:01.586 --rc genhtml_function_coverage=1 00:07:01.586 --rc genhtml_legend=1 00:07:01.586 --rc geninfo_all_blocks=1 00:07:01.586 --rc geninfo_unexecuted_blocks=1 00:07:01.586 00:07:01.586 ' 00:07:01.586 15:11:29 bdev_raid -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:07:01.586 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:01.586 --rc genhtml_branch_coverage=1 00:07:01.586 --rc genhtml_function_coverage=1 00:07:01.586 --rc genhtml_legend=1 00:07:01.586 --rc geninfo_all_blocks=1 00:07:01.586 --rc geninfo_unexecuted_blocks=1 00:07:01.586 00:07:01.586 ' 00:07:01.586 15:11:29 bdev_raid -- bdev/bdev_raid.sh@12 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:07:01.586 15:11:29 bdev_raid -- bdev/nbd_common.sh@6 -- # set -e 00:07:01.586 15:11:29 bdev_raid -- bdev/bdev_raid.sh@14 -- # rpc_py=rpc_cmd 00:07:01.586 15:11:29 bdev_raid -- bdev/bdev_raid.sh@946 -- # mkdir -p /raidtest 00:07:01.586 15:11:29 bdev_raid -- bdev/bdev_raid.sh@947 -- # trap 'cleanup; exit 1' EXIT 00:07:01.586 15:11:29 bdev_raid -- bdev/bdev_raid.sh@949 -- # base_blocklen=512 00:07:01.586 15:11:29 bdev_raid -- bdev/bdev_raid.sh@951 -- # run_test raid1_resize_data_offset_test raid_resize_data_offset_test 00:07:01.586 15:11:29 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:01.586 15:11:29 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:01.586 15:11:29 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:01.586 ************************************ 00:07:01.586 START TEST raid1_resize_data_offset_test 00:07:01.586 ************************************ 00:07:01.586 15:11:29 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@1129 -- # raid_resize_data_offset_test 00:07:01.586 15:11:29 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@917 -- # raid_pid=71755 00:07:01.586 Process raid pid: 71755 00:07:01.586 15:11:29 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@918 -- # echo 'Process raid pid: 71755' 00:07:01.586 15:11:29 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@916 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:01.586 15:11:29 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@919 -- # waitforlisten 71755 00:07:01.586 15:11:29 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@835 -- # '[' -z 71755 ']' 00:07:01.586 15:11:29 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:01.586 15:11:29 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:01.586 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:01.586 15:11:29 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:01.586 15:11:29 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:01.586 15:11:29 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:01.586 [2024-11-27 15:11:29.638002] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:07:01.586 [2024-11-27 15:11:29.638141] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:01.847 [2024-11-27 15:11:29.807989] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:01.847 [2024-11-27 15:11:29.833258] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:01.847 [2024-11-27 15:11:29.874641] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:01.847 [2024-11-27 15:11:29.874686] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:02.420 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:02.420 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@868 -- # return 0 00:07:02.420 15:11:30 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@922 -- # rpc_cmd bdev_malloc_create -b malloc0 64 512 -o 16 00:07:02.420 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:02.420 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:02.420 malloc0 00:07:02.420 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:02.420 15:11:30 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@923 -- # rpc_cmd bdev_malloc_create -b malloc1 64 512 -o 16 00:07:02.420 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:02.420 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:02.685 malloc1 00:07:02.685 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:02.685 15:11:30 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@924 -- # rpc_cmd bdev_null_create null0 64 512 00:07:02.685 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:02.685 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:02.685 null0 00:07:02.685 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:02.685 15:11:30 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@926 -- # rpc_cmd bdev_raid_create -n Raid -r 1 -b ''\''malloc0 malloc1 null0'\''' -s 00:07:02.685 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:02.685 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:02.685 [2024-11-27 15:11:30.551012] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc0 is claimed 00:07:02.685 [2024-11-27 15:11:30.552998] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:07:02.685 [2024-11-27 15:11:30.553049] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev null0 is claimed 00:07:02.685 [2024-11-27 15:11:30.553174] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:07:02.685 [2024-11-27 15:11:30.553185] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 129024, blocklen 512 00:07:02.685 [2024-11-27 15:11:30.553476] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005a00 00:07:02.685 [2024-11-27 15:11:30.553640] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:07:02.685 [2024-11-27 15:11:30.553658] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000006280 00:07:02.685 [2024-11-27 15:11:30.553791] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:02.685 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:02.685 15:11:30 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:02.685 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:02.685 15:11:30 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # jq -r '.[].base_bdevs_list[2].data_offset' 00:07:02.685 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:02.685 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:02.685 15:11:30 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # (( 2048 == 2048 )) 00:07:02.685 15:11:30 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@931 -- # rpc_cmd bdev_null_delete null0 00:07:02.685 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:02.685 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:02.685 [2024-11-27 15:11:30.610874] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: null0 00:07:02.685 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:02.685 15:11:30 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@935 -- # rpc_cmd bdev_malloc_create -b malloc2 512 512 -o 30 00:07:02.685 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:02.685 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:02.685 malloc2 00:07:02.685 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:02.685 15:11:30 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@936 -- # rpc_cmd bdev_raid_add_base_bdev Raid malloc2 00:07:02.685 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:02.685 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:02.685 [2024-11-27 15:11:30.733788] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:07:02.685 [2024-11-27 15:11:30.738782] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:07:02.685 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:02.685 [2024-11-27 15:11:30.740659] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev Raid 00:07:02.685 15:11:30 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:02.685 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:02.685 15:11:30 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # jq -r '.[].base_bdevs_list[2].data_offset' 00:07:02.685 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:02.685 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:02.944 15:11:30 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # (( 2070 == 2070 )) 00:07:02.944 15:11:30 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@941 -- # killprocess 71755 00:07:02.944 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@954 -- # '[' -z 71755 ']' 00:07:02.944 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@958 -- # kill -0 71755 00:07:02.944 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@959 -- # uname 00:07:02.944 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:02.944 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71755 00:07:02.944 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:02.944 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:02.944 killing process with pid 71755 00:07:02.944 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71755' 00:07:02.944 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@973 -- # kill 71755 00:07:02.944 [2024-11-27 15:11:30.820951] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:02.944 15:11:30 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@978 -- # wait 71755 00:07:02.944 [2024-11-27 15:11:30.822659] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev Raid: Operation canceled 00:07:02.944 [2024-11-27 15:11:30.822733] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:02.944 [2024-11-27 15:11:30.822750] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: malloc2 00:07:02.944 [2024-11-27 15:11:30.828980] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:02.944 [2024-11-27 15:11:30.829269] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:02.944 [2024-11-27 15:11:30.829290] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Raid, state offline 00:07:02.944 [2024-11-27 15:11:31.038195] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:03.203 15:11:31 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@943 -- # return 0 00:07:03.203 00:07:03.203 real 0m1.690s 00:07:03.203 user 0m1.712s 00:07:03.203 sys 0m0.430s 00:07:03.203 15:11:31 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:03.203 15:11:31 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:03.203 ************************************ 00:07:03.203 END TEST raid1_resize_data_offset_test 00:07:03.203 ************************************ 00:07:03.203 15:11:31 bdev_raid -- bdev/bdev_raid.sh@953 -- # run_test raid0_resize_superblock_test raid_resize_superblock_test 0 00:07:03.203 15:11:31 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:07:03.203 15:11:31 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:03.203 15:11:31 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:03.462 ************************************ 00:07:03.462 START TEST raid0_resize_superblock_test 00:07:03.462 ************************************ 00:07:03.462 15:11:31 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@1129 -- # raid_resize_superblock_test 0 00:07:03.462 15:11:31 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@854 -- # local raid_level=0 00:07:03.462 15:11:31 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@857 -- # raid_pid=71805 00:07:03.462 15:11:31 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@856 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:03.462 15:11:31 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@858 -- # echo 'Process raid pid: 71805' 00:07:03.462 Process raid pid: 71805 00:07:03.462 15:11:31 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@859 -- # waitforlisten 71805 00:07:03.462 15:11:31 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 71805 ']' 00:07:03.462 15:11:31 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:03.462 15:11:31 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:03.462 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:03.462 15:11:31 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:03.462 15:11:31 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:03.462 15:11:31 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:03.462 [2024-11-27 15:11:31.397717] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:07:03.462 [2024-11-27 15:11:31.397849] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:03.721 [2024-11-27 15:11:31.569728] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:03.721 [2024-11-27 15:11:31.596097] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:03.721 [2024-11-27 15:11:31.638682] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:03.721 [2024-11-27 15:11:31.638723] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:04.293 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:04.293 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:07:04.293 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@861 -- # rpc_cmd bdev_malloc_create -b malloc0 512 512 00:07:04.293 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.293 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.293 malloc0 00:07:04.293 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.293 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@863 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:07:04.293 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.293 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.293 [2024-11-27 15:11:32.348970] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:07:04.293 [2024-11-27 15:11:32.349052] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:04.293 [2024-11-27 15:11:32.349080] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:07:04.293 [2024-11-27 15:11:32.349092] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:04.294 [2024-11-27 15:11:32.351325] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:04.294 [2024-11-27 15:11:32.351364] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:07:04.294 pt0 00:07:04.294 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.294 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@864 -- # rpc_cmd bdev_lvol_create_lvstore pt0 lvs0 00:07:04.294 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.294 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.553 a4dd6bbf-771d-456c-a996-9cdd099cce90 00:07:04.553 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.553 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@866 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol0 64 00:07:04.553 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.553 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.553 33732d9f-5be7-48d7-af90-6bbb97ecaeb0 00:07:04.553 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.553 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@867 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol1 64 00:07:04.553 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.553 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.553 73cf0b0b-036c-4c90-9258-f2b20efd700f 00:07:04.553 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.553 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@869 -- # case $raid_level in 00:07:04.553 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@870 -- # rpc_cmd bdev_raid_create -n Raid -r 0 -z 64 -b ''\''lvs0/lvol0 lvs0/lvol1'\''' -s 00:07:04.553 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.553 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.553 [2024-11-27 15:11:32.485622] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 33732d9f-5be7-48d7-af90-6bbb97ecaeb0 is claimed 00:07:04.553 [2024-11-27 15:11:32.485737] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 73cf0b0b-036c-4c90-9258-f2b20efd700f is claimed 00:07:04.553 [2024-11-27 15:11:32.485851] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:07:04.553 [2024-11-27 15:11:32.485871] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 245760, blocklen 512 00:07:04.554 [2024-11-27 15:11:32.486145] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:07:04.554 [2024-11-27 15:11:32.486350] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:07:04.554 [2024-11-27 15:11:32.486380] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000006280 00:07:04.554 [2024-11-27 15:11:32.486522] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:04.554 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.554 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:07:04.554 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # jq '.[].num_blocks' 00:07:04.554 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.554 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.554 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.554 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # (( 64 == 64 )) 00:07:04.554 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:07:04.554 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # jq '.[].num_blocks' 00:07:04.554 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.554 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.554 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.554 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # (( 64 == 64 )) 00:07:04.554 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:07:04.554 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:04.554 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.554 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.554 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:07:04.554 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # jq '.[].num_blocks' 00:07:04.554 [2024-11-27 15:11:32.597646] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:04.554 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.554 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:07:04.554 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:07:04.554 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # (( 245760 == 245760 )) 00:07:04.554 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@885 -- # rpc_cmd bdev_lvol_resize lvs0/lvol0 100 00:07:04.554 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.554 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.554 [2024-11-27 15:11:32.645567] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:07:04.554 [2024-11-27 15:11:32.645602] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev '33732d9f-5be7-48d7-af90-6bbb97ecaeb0' was resized: old size 131072, new size 204800 00:07:04.554 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.554 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@886 -- # rpc_cmd bdev_lvol_resize lvs0/lvol1 100 00:07:04.554 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.554 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.554 [2024-11-27 15:11:32.657410] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:07:04.554 [2024-11-27 15:11:32.657437] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev '73cf0b0b-036c-4c90-9258-f2b20efd700f' was resized: old size 131072, new size 204800 00:07:04.554 [2024-11-27 15:11:32.657461] bdev_raid.c:2344:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 245760 to 393216 00:07:04.814 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.814 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # jq '.[].num_blocks' 00:07:04.814 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:07:04.814 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.814 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.814 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.814 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # (( 100 == 100 )) 00:07:04.814 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # jq '.[].num_blocks' 00:07:04.814 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:07:04.814 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.814 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.814 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.814 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # (( 100 == 100 )) 00:07:04.814 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:07:04.814 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:04.814 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.814 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.814 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:07:04.814 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # jq '.[].num_blocks' 00:07:04.814 [2024-11-27 15:11:32.757327] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:04.814 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.814 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:07:04.814 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:07:04.814 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # (( 393216 == 393216 )) 00:07:04.814 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@898 -- # rpc_cmd bdev_passthru_delete pt0 00:07:04.814 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.814 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.814 [2024-11-27 15:11:32.805077] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev pt0 being removed: closing lvstore lvs0 00:07:04.814 [2024-11-27 15:11:32.805150] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol0 00:07:04.815 [2024-11-27 15:11:32.805162] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:04.815 [2024-11-27 15:11:32.805184] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol1 00:07:04.815 [2024-11-27 15:11:32.805304] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:04.815 [2024-11-27 15:11:32.805339] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:04.815 [2024-11-27 15:11:32.805352] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Raid, state offline 00:07:04.815 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.815 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@899 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:07:04.815 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.815 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.815 [2024-11-27 15:11:32.817033] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:07:04.815 [2024-11-27 15:11:32.817082] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:04.815 [2024-11-27 15:11:32.817100] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008180 00:07:04.815 [2024-11-27 15:11:32.817111] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:04.815 [2024-11-27 15:11:32.819379] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:04.815 [2024-11-27 15:11:32.819416] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:07:04.815 [2024-11-27 15:11:32.821024] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev 33732d9f-5be7-48d7-af90-6bbb97ecaeb0 00:07:04.815 [2024-11-27 15:11:32.821077] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 33732d9f-5be7-48d7-af90-6bbb97ecaeb0 is claimed 00:07:04.815 [2024-11-27 15:11:32.821157] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev 73cf0b0b-036c-4c90-9258-f2b20efd700f 00:07:04.815 [2024-11-27 15:11:32.821179] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 73cf0b0b-036c-4c90-9258-f2b20efd700f is claimed 00:07:04.815 [2024-11-27 15:11:32.821291] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev 73cf0b0b-036c-4c90-9258-f2b20efd700f (2) smaller than existing raid bdev Raid (3) 00:07:04.815 [2024-11-27 15:11:32.821318] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev 33732d9f-5be7-48d7-af90-6bbb97ecaeb0: File exists 00:07:04.815 [2024-11-27 15:11:32.821380] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006600 00:07:04.815 [2024-11-27 15:11:32.821393] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 393216, blocklen 512 00:07:04.815 [2024-11-27 15:11:32.821643] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:07:04.815 pt0 00:07:04.815 [2024-11-27 15:11:32.821796] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006600 00:07:04.815 [2024-11-27 15:11:32.821807] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000006600 00:07:04.815 [2024-11-27 15:11:32.821952] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:04.815 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.815 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@900 -- # rpc_cmd bdev_wait_for_examine 00:07:04.815 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.815 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.815 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.815 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:07:04.815 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:04.815 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.815 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.815 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:07:04.815 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # jq '.[].num_blocks' 00:07:04.815 [2024-11-27 15:11:32.841314] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:04.815 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.815 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:07:04.815 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:07:04.815 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # (( 393216 == 393216 )) 00:07:04.815 15:11:32 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@909 -- # killprocess 71805 00:07:04.815 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 71805 ']' 00:07:04.815 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@958 -- # kill -0 71805 00:07:04.815 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@959 -- # uname 00:07:04.815 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:04.815 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71805 00:07:04.815 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:04.815 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:04.815 killing process with pid 71805 00:07:04.815 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71805' 00:07:04.815 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@973 -- # kill 71805 00:07:04.815 [2024-11-27 15:11:32.915837] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:04.815 [2024-11-27 15:11:32.915930] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:04.815 [2024-11-27 15:11:32.915977] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:04.815 [2024-11-27 15:11:32.915986] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Raid, state offline 00:07:04.815 15:11:32 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@978 -- # wait 71805 00:07:05.075 [2024-11-27 15:11:33.076006] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:05.334 15:11:33 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@911 -- # return 0 00:07:05.334 00:07:05.334 real 0m1.979s 00:07:05.334 user 0m2.277s 00:07:05.334 sys 0m0.476s 00:07:05.334 15:11:33 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:05.334 15:11:33 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:05.334 ************************************ 00:07:05.334 END TEST raid0_resize_superblock_test 00:07:05.334 ************************************ 00:07:05.334 15:11:33 bdev_raid -- bdev/bdev_raid.sh@954 -- # run_test raid1_resize_superblock_test raid_resize_superblock_test 1 00:07:05.334 15:11:33 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:07:05.334 15:11:33 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:05.334 15:11:33 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:05.334 ************************************ 00:07:05.334 START TEST raid1_resize_superblock_test 00:07:05.334 ************************************ 00:07:05.334 15:11:33 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@1129 -- # raid_resize_superblock_test 1 00:07:05.334 15:11:33 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@854 -- # local raid_level=1 00:07:05.334 15:11:33 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@857 -- # raid_pid=71876 00:07:05.334 15:11:33 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@856 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:05.334 15:11:33 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@858 -- # echo 'Process raid pid: 71876' 00:07:05.334 Process raid pid: 71876 00:07:05.334 15:11:33 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@859 -- # waitforlisten 71876 00:07:05.334 15:11:33 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 71876 ']' 00:07:05.334 15:11:33 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:05.334 15:11:33 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:05.334 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:05.334 15:11:33 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:05.334 15:11:33 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:05.334 15:11:33 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:05.594 [2024-11-27 15:11:33.453505] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:07:05.594 [2024-11-27 15:11:33.453652] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:05.594 [2024-11-27 15:11:33.625967] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:05.594 [2024-11-27 15:11:33.651059] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:05.594 [2024-11-27 15:11:33.692952] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:05.594 [2024-11-27 15:11:33.692991] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:06.532 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:06.532 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:07:06.532 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@861 -- # rpc_cmd bdev_malloc_create -b malloc0 512 512 00:07:06.532 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.532 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.532 malloc0 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@863 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.533 [2024-11-27 15:11:34.394525] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:07:06.533 [2024-11-27 15:11:34.394585] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:06.533 [2024-11-27 15:11:34.394623] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:07:06.533 [2024-11-27 15:11:34.394634] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:06.533 [2024-11-27 15:11:34.396825] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:06.533 [2024-11-27 15:11:34.396865] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:07:06.533 pt0 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@864 -- # rpc_cmd bdev_lvol_create_lvstore pt0 lvs0 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.533 ea2168c1-2f4b-4ddb-87a4-61c10c2d047c 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@866 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol0 64 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.533 32afa3f0-890a-4239-a233-1d07a940c252 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@867 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol1 64 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.533 2d1588f5-99e3-4fab-8d63-bf5f294c8bcb 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@869 -- # case $raid_level in 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@871 -- # rpc_cmd bdev_raid_create -n Raid -r 1 -b ''\''lvs0/lvol0 lvs0/lvol1'\''' -s 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.533 [2024-11-27 15:11:34.530286] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 32afa3f0-890a-4239-a233-1d07a940c252 is claimed 00:07:06.533 [2024-11-27 15:11:34.530373] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 2d1588f5-99e3-4fab-8d63-bf5f294c8bcb is claimed 00:07:06.533 [2024-11-27 15:11:34.530491] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:07:06.533 [2024-11-27 15:11:34.530513] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 122880, blocklen 512 00:07:06.533 [2024-11-27 15:11:34.530748] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:07:06.533 [2024-11-27 15:11:34.530947] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:07:06.533 [2024-11-27 15:11:34.530964] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000006280 00:07:06.533 [2024-11-27 15:11:34.531089] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # jq '.[].num_blocks' 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # (( 64 == 64 )) 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # jq '.[].num_blocks' 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # (( 64 == 64 )) 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # jq '.[].num_blocks' 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.533 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.533 [2024-11-27 15:11:34.622308] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:06.792 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.792 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:07:06.792 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:07:06.792 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # (( 122880 == 122880 )) 00:07:06.792 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@885 -- # rpc_cmd bdev_lvol_resize lvs0/lvol0 100 00:07:06.792 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.792 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.792 [2024-11-27 15:11:34.650248] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:07:06.792 [2024-11-27 15:11:34.650281] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev '32afa3f0-890a-4239-a233-1d07a940c252' was resized: old size 131072, new size 204800 00:07:06.792 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.792 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@886 -- # rpc_cmd bdev_lvol_resize lvs0/lvol1 100 00:07:06.792 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.792 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.792 [2024-11-27 15:11:34.662122] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:07:06.792 [2024-11-27 15:11:34.662149] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev '2d1588f5-99e3-4fab-8d63-bf5f294c8bcb' was resized: old size 131072, new size 204800 00:07:06.792 [2024-11-27 15:11:34.662179] bdev_raid.c:2344:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 122880 to 196608 00:07:06.792 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.792 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:07:06.792 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # jq '.[].num_blocks' 00:07:06.792 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.792 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.792 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.792 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # (( 100 == 100 )) 00:07:06.792 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:07:06.792 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # jq '.[].num_blocks' 00:07:06.792 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.792 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.792 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.792 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # (( 100 == 100 )) 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # jq '.[].num_blocks' 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.793 [2024-11-27 15:11:34.774029] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # (( 196608 == 196608 )) 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@898 -- # rpc_cmd bdev_passthru_delete pt0 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.793 [2024-11-27 15:11:34.817759] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev pt0 being removed: closing lvstore lvs0 00:07:06.793 [2024-11-27 15:11:34.817842] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol0 00:07:06.793 [2024-11-27 15:11:34.817870] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol1 00:07:06.793 [2024-11-27 15:11:34.818045] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:06.793 [2024-11-27 15:11:34.818227] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:06.793 [2024-11-27 15:11:34.818296] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:06.793 [2024-11-27 15:11:34.818311] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Raid, state offline 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@899 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.793 [2024-11-27 15:11:34.829686] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:07:06.793 [2024-11-27 15:11:34.829732] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:06.793 [2024-11-27 15:11:34.829749] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008180 00:07:06.793 [2024-11-27 15:11:34.829776] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:06.793 [2024-11-27 15:11:34.831898] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:06.793 [2024-11-27 15:11:34.831946] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:07:06.793 [2024-11-27 15:11:34.833501] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev 32afa3f0-890a-4239-a233-1d07a940c252 00:07:06.793 [2024-11-27 15:11:34.833568] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 32afa3f0-890a-4239-a233-1d07a940c252 is claimed 00:07:06.793 [2024-11-27 15:11:34.833639] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev 2d1588f5-99e3-4fab-8d63-bf5f294c8bcb 00:07:06.793 [2024-11-27 15:11:34.833659] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 2d1588f5-99e3-4fab-8d63-bf5f294c8bcb is claimed 00:07:06.793 [2024-11-27 15:11:34.833736] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev 2d1588f5-99e3-4fab-8d63-bf5f294c8bcb (2) smaller than existing raid bdev Raid (3) 00:07:06.793 [2024-11-27 15:11:34.833760] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev 32afa3f0-890a-4239-a233-1d07a940c252: File exists 00:07:06.793 [2024-11-27 15:11:34.833814] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006600 00:07:06.793 [2024-11-27 15:11:34.833835] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:07:06.793 [2024-11-27 15:11:34.834088] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:07:06.793 [2024-11-27 15:11:34.834237] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006600 00:07:06.793 [2024-11-27 15:11:34.834251] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000006600 00:07:06.793 [2024-11-27 15:11:34.834394] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:06.793 pt0 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@900 -- # rpc_cmd bdev_wait_for_examine 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # jq '.[].num_blocks' 00:07:06.793 [2024-11-27 15:11:34.854017] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # (( 196608 == 196608 )) 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@909 -- # killprocess 71876 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 71876 ']' 00:07:06.793 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@958 -- # kill -0 71876 00:07:07.052 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@959 -- # uname 00:07:07.052 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:07.052 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71876 00:07:07.052 killing process with pid 71876 00:07:07.052 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:07.052 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:07.052 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71876' 00:07:07.052 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@973 -- # kill 71876 00:07:07.052 [2024-11-27 15:11:34.937256] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:07.052 [2024-11-27 15:11:34.937362] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:07.052 15:11:34 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@978 -- # wait 71876 00:07:07.052 [2024-11-27 15:11:34.937423] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:07.052 [2024-11-27 15:11:34.937435] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Raid, state offline 00:07:07.052 [2024-11-27 15:11:35.099254] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:07.313 15:11:35 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@911 -- # return 0 00:07:07.313 00:07:07.313 real 0m1.951s 00:07:07.313 user 0m2.198s 00:07:07.313 sys 0m0.500s 00:07:07.313 15:11:35 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:07.313 15:11:35 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:07.313 ************************************ 00:07:07.313 END TEST raid1_resize_superblock_test 00:07:07.313 ************************************ 00:07:07.313 15:11:35 bdev_raid -- bdev/bdev_raid.sh@956 -- # uname -s 00:07:07.313 15:11:35 bdev_raid -- bdev/bdev_raid.sh@956 -- # '[' Linux = Linux ']' 00:07:07.313 15:11:35 bdev_raid -- bdev/bdev_raid.sh@956 -- # modprobe -n nbd 00:07:07.313 15:11:35 bdev_raid -- bdev/bdev_raid.sh@957 -- # has_nbd=true 00:07:07.313 15:11:35 bdev_raid -- bdev/bdev_raid.sh@958 -- # modprobe nbd 00:07:07.313 15:11:35 bdev_raid -- bdev/bdev_raid.sh@959 -- # run_test raid_function_test_raid0 raid_function_test raid0 00:07:07.313 15:11:35 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:07:07.313 15:11:35 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:07.313 15:11:35 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:07.313 ************************************ 00:07:07.313 START TEST raid_function_test_raid0 00:07:07.313 ************************************ 00:07:07.313 15:11:35 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@1129 -- # raid_function_test raid0 00:07:07.573 15:11:35 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@64 -- # local raid_level=raid0 00:07:07.573 15:11:35 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@65 -- # local nbd=/dev/nbd0 00:07:07.573 15:11:35 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@66 -- # local raid_bdev 00:07:07.573 15:11:35 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@69 -- # raid_pid=71951 00:07:07.573 15:11:35 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@68 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:07.573 Process raid pid: 71951 00:07:07.573 15:11:35 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@70 -- # echo 'Process raid pid: 71951' 00:07:07.573 15:11:35 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@71 -- # waitforlisten 71951 00:07:07.573 15:11:35 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@835 -- # '[' -z 71951 ']' 00:07:07.573 15:11:35 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:07.573 15:11:35 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:07.573 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:07.573 15:11:35 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:07.573 15:11:35 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:07.573 15:11:35 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:07:07.573 [2024-11-27 15:11:35.517402] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:07:07.573 [2024-11-27 15:11:35.517595] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:07.833 [2024-11-27 15:11:35.694982] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:07.833 [2024-11-27 15:11:35.722548] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:07.833 [2024-11-27 15:11:35.766098] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:07.833 [2024-11-27 15:11:35.766141] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:08.401 15:11:36 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:08.401 15:11:36 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@868 -- # return 0 00:07:08.401 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@73 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_1 00:07:08.401 15:11:36 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.401 15:11:36 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:07:08.401 Base_1 00:07:08.401 15:11:36 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.401 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@74 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_2 00:07:08.401 15:11:36 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.401 15:11:36 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:07:08.401 Base_2 00:07:08.401 15:11:36 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.401 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@75 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''Base_1 Base_2'\''' -n raid 00:07:08.401 15:11:36 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.401 15:11:36 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:07:08.401 [2024-11-27 15:11:36.375871] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:07:08.401 [2024-11-27 15:11:36.377724] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:07:08.401 [2024-11-27 15:11:36.377791] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:07:08.401 [2024-11-27 15:11:36.377811] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:07:08.401 [2024-11-27 15:11:36.378106] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:07:08.401 [2024-11-27 15:11:36.378249] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:07:08.401 [2024-11-27 15:11:36.378264] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid, raid_bdev 0x617000006280 00:07:08.401 [2024-11-27 15:11:36.378411] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:08.401 15:11:36 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.401 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # rpc_cmd bdev_raid_get_bdevs online 00:07:08.401 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # jq -r '.[0]["name"] | select(.)' 00:07:08.401 15:11:36 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.401 15:11:36 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:07:08.401 15:11:36 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.401 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # raid_bdev=raid 00:07:08.401 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@78 -- # '[' raid = '' ']' 00:07:08.401 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@83 -- # nbd_start_disks /var/tmp/spdk.sock raid /dev/nbd0 00:07:08.401 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:07:08.401 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@10 -- # bdev_list=('raid') 00:07:08.401 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@10 -- # local bdev_list 00:07:08.401 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:07:08.401 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@11 -- # local nbd_list 00:07:08.401 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@12 -- # local i 00:07:08.401 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:07:08.401 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:07:08.401 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid /dev/nbd0 00:07:08.661 [2024-11-27 15:11:36.615545] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:07:08.661 /dev/nbd0 00:07:08.661 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:07:08.661 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:07:08.661 15:11:36 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:07:08.661 15:11:36 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@873 -- # local i 00:07:08.661 15:11:36 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:07:08.661 15:11:36 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:07:08.661 15:11:36 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:07:08.661 15:11:36 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@877 -- # break 00:07:08.661 15:11:36 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:07:08.661 15:11:36 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:07:08.661 15:11:36 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:07:08.661 1+0 records in 00:07:08.661 1+0 records out 00:07:08.661 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000367914 s, 11.1 MB/s 00:07:08.661 15:11:36 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:07:08.661 15:11:36 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@890 -- # size=4096 00:07:08.661 15:11:36 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:07:08.661 15:11:36 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:07:08.661 15:11:36 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@893 -- # return 0 00:07:08.661 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:07:08.661 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:07:08.661 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@84 -- # nbd_get_count /var/tmp/spdk.sock 00:07:08.661 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:07:08.662 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:07:08.921 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:07:08.921 { 00:07:08.921 "nbd_device": "/dev/nbd0", 00:07:08.921 "bdev_name": "raid" 00:07:08.921 } 00:07:08.921 ]' 00:07:08.921 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # echo '[ 00:07:08.921 { 00:07:08.921 "nbd_device": "/dev/nbd0", 00:07:08.921 "bdev_name": "raid" 00:07:08.921 } 00:07:08.921 ]' 00:07:08.921 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:08.921 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:07:08.921 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:07:08.921 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:08.921 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # count=1 00:07:08.921 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@66 -- # echo 1 00:07:08.921 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@84 -- # count=1 00:07:08.921 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@85 -- # '[' 1 -ne 1 ']' 00:07:08.921 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@89 -- # raid_unmap_data_verify /dev/nbd0 00:07:08.921 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@17 -- # hash blkdiscard 00:07:08.921 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@18 -- # local nbd=/dev/nbd0 00:07:08.922 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@19 -- # local blksize 00:07:08.922 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # cut -d ' ' -f 5 00:07:08.922 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # lsblk -o LOG-SEC /dev/nbd0 00:07:08.922 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # grep -v LOG-SEC 00:07:08.922 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # blksize=512 00:07:08.922 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@21 -- # local rw_blk_num=4096 00:07:08.922 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@22 -- # local rw_len=2097152 00:07:08.922 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@23 -- # unmap_blk_offs=('0' '1028' '321') 00:07:08.922 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@23 -- # local unmap_blk_offs 00:07:08.922 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@24 -- # unmap_blk_nums=('128' '2035' '456') 00:07:08.922 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@24 -- # local unmap_blk_nums 00:07:08.922 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@25 -- # local unmap_off 00:07:08.922 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@26 -- # local unmap_len 00:07:08.922 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@29 -- # dd if=/dev/urandom of=/raidtest/raidrandtest bs=512 count=4096 00:07:08.922 4096+0 records in 00:07:08.922 4096+0 records out 00:07:08.922 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.0315354 s, 66.5 MB/s 00:07:08.922 15:11:36 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@30 -- # dd if=/raidtest/raidrandtest of=/dev/nbd0 bs=512 count=4096 oflag=direct 00:07:09.181 4096+0 records in 00:07:09.181 4096+0 records out 00:07:09.181 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.185698 s, 11.3 MB/s 00:07:09.181 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@31 -- # blockdev --flushbufs /dev/nbd0 00:07:09.181 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@34 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:07:09.181 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i = 0 )) 00:07:09.181 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:07:09.181 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=0 00:07:09.181 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=65536 00:07:09.181 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=0 count=128 conv=notrunc 00:07:09.181 128+0 records in 00:07:09.181 128+0 records out 00:07:09.181 65536 bytes (66 kB, 64 KiB) copied, 0.0011968 s, 54.8 MB/s 00:07:09.181 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 0 -l 65536 /dev/nbd0 00:07:09.181 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:07:09.181 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:07:09.181 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:07:09.181 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:07:09.181 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=526336 00:07:09.181 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=1041920 00:07:09.181 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=1028 count=2035 conv=notrunc 00:07:09.181 2035+0 records in 00:07:09.181 2035+0 records out 00:07:09.181 1041920 bytes (1.0 MB, 1018 KiB) copied, 0.0147041 s, 70.9 MB/s 00:07:09.181 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 526336 -l 1041920 /dev/nbd0 00:07:09.181 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:07:09.181 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:07:09.181 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:07:09.181 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:07:09.181 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=164352 00:07:09.181 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=233472 00:07:09.181 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=321 count=456 conv=notrunc 00:07:09.181 456+0 records in 00:07:09.181 456+0 records out 00:07:09.181 233472 bytes (233 kB, 228 KiB) copied, 0.00387036 s, 60.3 MB/s 00:07:09.181 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 164352 -l 233472 /dev/nbd0 00:07:09.181 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:07:09.181 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:07:09.440 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:07:09.440 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:07:09.440 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@52 -- # return 0 00:07:09.440 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@91 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:07:09.440 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:07:09.440 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:07:09.440 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@50 -- # local nbd_list 00:07:09.440 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@51 -- # local i 00:07:09.440 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:07:09.440 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:07:09.440 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:07:09.440 [2024-11-27 15:11:37.495279] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:09.440 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:07:09.440 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:07:09.440 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:07:09.440 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:07:09.440 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:07:09.440 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@41 -- # break 00:07:09.440 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@45 -- # return 0 00:07:09.440 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@92 -- # nbd_get_count /var/tmp/spdk.sock 00:07:09.440 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:07:09.440 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:07:09.699 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:07:09.699 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # echo '[]' 00:07:09.699 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:09.699 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:07:09.699 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # echo '' 00:07:09.699 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:09.699 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # true 00:07:09.699 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # count=0 00:07:09.699 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@66 -- # echo 0 00:07:09.699 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@92 -- # count=0 00:07:09.699 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@93 -- # '[' 0 -ne 0 ']' 00:07:09.699 15:11:37 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@97 -- # killprocess 71951 00:07:09.699 15:11:37 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@954 -- # '[' -z 71951 ']' 00:07:09.699 15:11:37 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@958 -- # kill -0 71951 00:07:09.699 15:11:37 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@959 -- # uname 00:07:09.699 15:11:37 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:09.699 15:11:37 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71951 00:07:09.699 15:11:37 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:09.699 15:11:37 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:09.699 killing process with pid 71951 00:07:09.699 15:11:37 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71951' 00:07:09.700 15:11:37 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@973 -- # kill 71951 00:07:09.700 [2024-11-27 15:11:37.800275] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:09.700 [2024-11-27 15:11:37.800401] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:09.700 15:11:37 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@978 -- # wait 71951 00:07:09.700 [2024-11-27 15:11:37.800455] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:09.700 [2024-11-27 15:11:37.800466] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid, state offline 00:07:09.959 [2024-11-27 15:11:37.824122] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:09.959 15:11:38 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@99 -- # return 0 00:07:09.959 00:07:09.959 real 0m2.627s 00:07:09.959 user 0m3.271s 00:07:09.959 sys 0m0.897s 00:07:09.959 15:11:38 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:09.959 15:11:38 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:07:09.959 ************************************ 00:07:09.959 END TEST raid_function_test_raid0 00:07:09.959 ************************************ 00:07:10.218 15:11:38 bdev_raid -- bdev/bdev_raid.sh@960 -- # run_test raid_function_test_concat raid_function_test concat 00:07:10.218 15:11:38 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:07:10.218 15:11:38 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:10.218 15:11:38 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:10.218 ************************************ 00:07:10.218 START TEST raid_function_test_concat 00:07:10.218 ************************************ 00:07:10.218 15:11:38 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@1129 -- # raid_function_test concat 00:07:10.218 15:11:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@64 -- # local raid_level=concat 00:07:10.219 15:11:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@65 -- # local nbd=/dev/nbd0 00:07:10.219 15:11:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@66 -- # local raid_bdev 00:07:10.219 15:11:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@69 -- # raid_pid=72071 00:07:10.219 15:11:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@68 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:10.219 15:11:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@70 -- # echo 'Process raid pid: 72071' 00:07:10.219 Process raid pid: 72071 00:07:10.219 15:11:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@71 -- # waitforlisten 72071 00:07:10.219 15:11:38 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@835 -- # '[' -z 72071 ']' 00:07:10.219 15:11:38 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:10.219 15:11:38 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:10.219 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:10.219 15:11:38 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:10.219 15:11:38 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:10.219 15:11:38 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:07:10.219 [2024-11-27 15:11:38.190081] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:07:10.219 [2024-11-27 15:11:38.190213] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:10.478 [2024-11-27 15:11:38.358822] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:10.478 [2024-11-27 15:11:38.384232] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:10.478 [2024-11-27 15:11:38.427199] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:10.478 [2024-11-27 15:11:38.427238] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:11.044 15:11:39 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:11.044 15:11:39 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@868 -- # return 0 00:07:11.044 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@73 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_1 00:07:11.044 15:11:39 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:11.044 15:11:39 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:07:11.044 Base_1 00:07:11.044 15:11:39 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:11.044 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@74 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_2 00:07:11.044 15:11:39 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:11.044 15:11:39 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:07:11.044 Base_2 00:07:11.044 15:11:39 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:11.044 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@75 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''Base_1 Base_2'\''' -n raid 00:07:11.044 15:11:39 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:11.044 15:11:39 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:07:11.044 [2024-11-27 15:11:39.060060] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:07:11.044 [2024-11-27 15:11:39.061891] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:07:11.044 [2024-11-27 15:11:39.061997] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:07:11.044 [2024-11-27 15:11:39.062010] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:07:11.045 [2024-11-27 15:11:39.062270] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:07:11.045 [2024-11-27 15:11:39.062419] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:07:11.045 [2024-11-27 15:11:39.062433] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid, raid_bdev 0x617000006280 00:07:11.045 [2024-11-27 15:11:39.062558] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:11.045 15:11:39 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:11.045 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # rpc_cmd bdev_raid_get_bdevs online 00:07:11.045 15:11:39 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:11.045 15:11:39 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:07:11.045 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # jq -r '.[0]["name"] | select(.)' 00:07:11.045 15:11:39 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:11.045 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # raid_bdev=raid 00:07:11.045 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@78 -- # '[' raid = '' ']' 00:07:11.045 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@83 -- # nbd_start_disks /var/tmp/spdk.sock raid /dev/nbd0 00:07:11.045 15:11:39 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:07:11.045 15:11:39 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@10 -- # bdev_list=('raid') 00:07:11.045 15:11:39 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:07:11.045 15:11:39 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:07:11.045 15:11:39 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:07:11.045 15:11:39 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@12 -- # local i 00:07:11.045 15:11:39 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:07:11.045 15:11:39 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:07:11.045 15:11:39 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid /dev/nbd0 00:07:11.304 [2024-11-27 15:11:39.299698] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:07:11.304 /dev/nbd0 00:07:11.304 15:11:39 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:07:11.304 15:11:39 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:07:11.304 15:11:39 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:07:11.304 15:11:39 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@873 -- # local i 00:07:11.304 15:11:39 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:07:11.304 15:11:39 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:07:11.304 15:11:39 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:07:11.304 15:11:39 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@877 -- # break 00:07:11.304 15:11:39 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:07:11.304 15:11:39 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:07:11.304 15:11:39 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:07:11.304 1+0 records in 00:07:11.304 1+0 records out 00:07:11.304 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000423949 s, 9.7 MB/s 00:07:11.304 15:11:39 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:07:11.304 15:11:39 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@890 -- # size=4096 00:07:11.304 15:11:39 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:07:11.304 15:11:39 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:07:11.304 15:11:39 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@893 -- # return 0 00:07:11.304 15:11:39 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:07:11.304 15:11:39 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:07:11.304 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@84 -- # nbd_get_count /var/tmp/spdk.sock 00:07:11.304 15:11:39 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:07:11.304 15:11:39 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:07:11.564 15:11:39 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:07:11.564 { 00:07:11.564 "nbd_device": "/dev/nbd0", 00:07:11.564 "bdev_name": "raid" 00:07:11.564 } 00:07:11.564 ]' 00:07:11.564 15:11:39 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # echo '[ 00:07:11.564 { 00:07:11.564 "nbd_device": "/dev/nbd0", 00:07:11.564 "bdev_name": "raid" 00:07:11.564 } 00:07:11.564 ]' 00:07:11.564 15:11:39 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:11.564 15:11:39 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:07:11.564 15:11:39 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:07:11.564 15:11:39 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:11.564 15:11:39 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # count=1 00:07:11.564 15:11:39 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@66 -- # echo 1 00:07:11.564 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@84 -- # count=1 00:07:11.564 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@85 -- # '[' 1 -ne 1 ']' 00:07:11.564 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@89 -- # raid_unmap_data_verify /dev/nbd0 00:07:11.564 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@17 -- # hash blkdiscard 00:07:11.564 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@18 -- # local nbd=/dev/nbd0 00:07:11.564 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@19 -- # local blksize 00:07:11.564 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # lsblk -o LOG-SEC /dev/nbd0 00:07:11.564 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # grep -v LOG-SEC 00:07:11.564 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # cut -d ' ' -f 5 00:07:11.564 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # blksize=512 00:07:11.564 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@21 -- # local rw_blk_num=4096 00:07:11.564 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@22 -- # local rw_len=2097152 00:07:11.564 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@23 -- # unmap_blk_offs=('0' '1028' '321') 00:07:11.564 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@23 -- # local unmap_blk_offs 00:07:11.564 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@24 -- # unmap_blk_nums=('128' '2035' '456') 00:07:11.564 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@24 -- # local unmap_blk_nums 00:07:11.564 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@25 -- # local unmap_off 00:07:11.564 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@26 -- # local unmap_len 00:07:11.564 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@29 -- # dd if=/dev/urandom of=/raidtest/raidrandtest bs=512 count=4096 00:07:11.824 4096+0 records in 00:07:11.824 4096+0 records out 00:07:11.824 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.0330897 s, 63.4 MB/s 00:07:11.824 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@30 -- # dd if=/raidtest/raidrandtest of=/dev/nbd0 bs=512 count=4096 oflag=direct 00:07:11.824 4096+0 records in 00:07:11.824 4096+0 records out 00:07:11.824 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.205376 s, 10.2 MB/s 00:07:11.824 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@31 -- # blockdev --flushbufs /dev/nbd0 00:07:11.824 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@34 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:07:11.824 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i = 0 )) 00:07:11.824 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:07:11.824 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=0 00:07:11.824 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=65536 00:07:11.824 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=0 count=128 conv=notrunc 00:07:11.824 128+0 records in 00:07:11.824 128+0 records out 00:07:11.824 65536 bytes (66 kB, 64 KiB) copied, 0.00109338 s, 59.9 MB/s 00:07:11.824 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 0 -l 65536 /dev/nbd0 00:07:11.824 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:07:11.824 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:07:11.824 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:07:11.824 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:07:11.824 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=526336 00:07:11.824 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=1041920 00:07:11.824 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=1028 count=2035 conv=notrunc 00:07:12.083 2035+0 records in 00:07:12.083 2035+0 records out 00:07:12.083 1041920 bytes (1.0 MB, 1018 KiB) copied, 0.012678 s, 82.2 MB/s 00:07:12.083 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 526336 -l 1041920 /dev/nbd0 00:07:12.083 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:07:12.083 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:07:12.083 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:07:12.083 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:07:12.083 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=164352 00:07:12.083 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=233472 00:07:12.083 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=321 count=456 conv=notrunc 00:07:12.083 456+0 records in 00:07:12.083 456+0 records out 00:07:12.083 233472 bytes (233 kB, 228 KiB) copied, 0.00347453 s, 67.2 MB/s 00:07:12.083 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 164352 -l 233472 /dev/nbd0 00:07:12.083 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:07:12.083 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:07:12.083 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:07:12.083 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:07:12.083 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@52 -- # return 0 00:07:12.083 15:11:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@91 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:07:12.083 15:11:39 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:07:12.083 15:11:39 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:07:12.083 15:11:39 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:07:12.083 15:11:39 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@51 -- # local i 00:07:12.083 15:11:39 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:07:12.083 15:11:39 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:07:12.342 15:11:40 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:07:12.342 [2024-11-27 15:11:40.195401] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:12.342 15:11:40 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:07:12.342 15:11:40 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:07:12.342 15:11:40 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:07:12.342 15:11:40 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:07:12.342 15:11:40 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:07:12.342 15:11:40 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@41 -- # break 00:07:12.342 15:11:40 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@45 -- # return 0 00:07:12.342 15:11:40 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@92 -- # nbd_get_count /var/tmp/spdk.sock 00:07:12.342 15:11:40 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:07:12.342 15:11:40 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:07:12.342 15:11:40 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:07:12.342 15:11:40 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:12.342 15:11:40 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:07:12.342 15:11:40 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:07:12.342 15:11:40 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # echo '' 00:07:12.342 15:11:40 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:12.342 15:11:40 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # true 00:07:12.342 15:11:40 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # count=0 00:07:12.342 15:11:40 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@66 -- # echo 0 00:07:12.342 15:11:40 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@92 -- # count=0 00:07:12.601 15:11:40 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@93 -- # '[' 0 -ne 0 ']' 00:07:12.601 15:11:40 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@97 -- # killprocess 72071 00:07:12.601 15:11:40 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@954 -- # '[' -z 72071 ']' 00:07:12.601 15:11:40 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@958 -- # kill -0 72071 00:07:12.601 15:11:40 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@959 -- # uname 00:07:12.601 15:11:40 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:12.601 15:11:40 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72071 00:07:12.601 15:11:40 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:12.601 15:11:40 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:12.601 killing process with pid 72071 00:07:12.601 15:11:40 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72071' 00:07:12.601 15:11:40 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@973 -- # kill 72071 00:07:12.601 [2024-11-27 15:11:40.484283] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:12.601 [2024-11-27 15:11:40.484401] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:12.601 15:11:40 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@978 -- # wait 72071 00:07:12.601 [2024-11-27 15:11:40.484463] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:12.601 [2024-11-27 15:11:40.484482] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid, state offline 00:07:12.601 [2024-11-27 15:11:40.508013] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:12.861 15:11:40 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@99 -- # return 0 00:07:12.861 00:07:12.861 real 0m2.618s 00:07:12.861 user 0m3.215s 00:07:12.861 sys 0m0.912s 00:07:12.861 15:11:40 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:12.861 15:11:40 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:07:12.861 ************************************ 00:07:12.861 END TEST raid_function_test_concat 00:07:12.861 ************************************ 00:07:12.861 15:11:40 bdev_raid -- bdev/bdev_raid.sh@963 -- # run_test raid0_resize_test raid_resize_test 0 00:07:12.861 15:11:40 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:07:12.861 15:11:40 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:12.861 15:11:40 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:12.861 ************************************ 00:07:12.861 START TEST raid0_resize_test 00:07:12.861 ************************************ 00:07:12.861 15:11:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@1129 -- # raid_resize_test 0 00:07:12.861 15:11:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@332 -- # local raid_level=0 00:07:12.861 15:11:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@333 -- # local blksize=512 00:07:12.861 15:11:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@334 -- # local bdev_size_mb=32 00:07:12.861 15:11:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@335 -- # local new_bdev_size_mb=64 00:07:12.861 15:11:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@336 -- # local blkcnt 00:07:12.861 15:11:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@337 -- # local raid_size_mb 00:07:12.861 15:11:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@338 -- # local new_raid_size_mb 00:07:12.861 15:11:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@339 -- # local expected_size 00:07:12.861 15:11:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@342 -- # raid_pid=72183 00:07:12.861 15:11:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@341 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:12.861 Process raid pid: 72183 00:07:12.861 15:11:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@343 -- # echo 'Process raid pid: 72183' 00:07:12.861 15:11:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@344 -- # waitforlisten 72183 00:07:12.861 15:11:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@835 -- # '[' -z 72183 ']' 00:07:12.861 15:11:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:12.861 15:11:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:12.861 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:12.861 15:11:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:12.861 15:11:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:12.861 15:11:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:12.861 [2024-11-27 15:11:40.876091] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:07:12.861 [2024-11-27 15:11:40.876244] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:13.125 [2024-11-27 15:11:41.046466] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:13.125 [2024-11-27 15:11:41.074139] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:13.125 [2024-11-27 15:11:41.117506] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:13.125 [2024-11-27 15:11:41.117558] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:13.701 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:13.701 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@868 -- # return 0 00:07:13.701 15:11:41 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@346 -- # rpc_cmd bdev_null_create Base_1 32 512 00:07:13.701 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:13.701 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.701 Base_1 00:07:13.701 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:13.701 15:11:41 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@347 -- # rpc_cmd bdev_null_create Base_2 32 512 00:07:13.701 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:13.701 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.701 Base_2 00:07:13.701 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:13.701 15:11:41 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@349 -- # '[' 0 -eq 0 ']' 00:07:13.701 15:11:41 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@350 -- # rpc_cmd bdev_raid_create -z 64 -r 0 -b ''\''Base_1 Base_2'\''' -n Raid 00:07:13.701 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:13.701 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.701 [2024-11-27 15:11:41.732836] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:07:13.701 [2024-11-27 15:11:41.734611] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:07:13.701 [2024-11-27 15:11:41.734666] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:07:13.701 [2024-11-27 15:11:41.734678] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:07:13.701 [2024-11-27 15:11:41.734953] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005a00 00:07:13.701 [2024-11-27 15:11:41.735075] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:07:13.701 [2024-11-27 15:11:41.735090] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000006280 00:07:13.701 [2024-11-27 15:11:41.735199] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:13.701 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:13.701 15:11:41 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@356 -- # rpc_cmd bdev_null_resize Base_1 64 00:07:13.701 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:13.701 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.701 [2024-11-27 15:11:41.744821] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:07:13.701 [2024-11-27 15:11:41.744850] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_1' was resized: old size 65536, new size 131072 00:07:13.701 true 00:07:13.701 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:13.701 15:11:41 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:13.701 15:11:41 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # jq '.[].num_blocks' 00:07:13.701 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:13.701 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.701 [2024-11-27 15:11:41.760978] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:13.701 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:13.701 15:11:41 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # blkcnt=131072 00:07:13.701 15:11:41 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@360 -- # raid_size_mb=64 00:07:13.701 15:11:41 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@361 -- # '[' 0 -eq 0 ']' 00:07:13.701 15:11:41 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@362 -- # expected_size=64 00:07:13.701 15:11:41 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@366 -- # '[' 64 '!=' 64 ']' 00:07:13.701 15:11:41 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@372 -- # rpc_cmd bdev_null_resize Base_2 64 00:07:13.701 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:13.701 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.701 [2024-11-27 15:11:41.804710] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:07:13.701 [2024-11-27 15:11:41.804739] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_2' was resized: old size 65536, new size 131072 00:07:13.701 [2024-11-27 15:11:41.804766] bdev_raid.c:2344:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 131072 to 262144 00:07:13.961 true 00:07:13.961 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:13.961 15:11:41 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:13.961 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:13.961 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.961 15:11:41 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # jq '.[].num_blocks' 00:07:13.961 [2024-11-27 15:11:41.816861] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:13.961 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:13.961 15:11:41 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # blkcnt=262144 00:07:13.961 15:11:41 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@376 -- # raid_size_mb=128 00:07:13.961 15:11:41 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@377 -- # '[' 0 -eq 0 ']' 00:07:13.961 15:11:41 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@378 -- # expected_size=128 00:07:13.961 15:11:41 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@382 -- # '[' 128 '!=' 128 ']' 00:07:13.961 15:11:41 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@387 -- # killprocess 72183 00:07:13.961 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@954 -- # '[' -z 72183 ']' 00:07:13.961 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@958 -- # kill -0 72183 00:07:13.961 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@959 -- # uname 00:07:13.961 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:13.961 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72183 00:07:13.961 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:13.961 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:13.961 killing process with pid 72183 00:07:13.961 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72183' 00:07:13.961 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@973 -- # kill 72183 00:07:13.961 [2024-11-27 15:11:41.905051] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:13.961 [2024-11-27 15:11:41.905134] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:13.961 [2024-11-27 15:11:41.905183] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:13.961 [2024-11-27 15:11:41.905192] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Raid, state offline 00:07:13.961 15:11:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@978 -- # wait 72183 00:07:13.961 [2024-11-27 15:11:41.906709] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:14.221 15:11:42 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@389 -- # return 0 00:07:14.221 00:07:14.221 real 0m1.330s 00:07:14.221 user 0m1.498s 00:07:14.221 sys 0m0.298s 00:07:14.221 15:11:42 bdev_raid.raid0_resize_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:14.221 15:11:42 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:14.221 ************************************ 00:07:14.221 END TEST raid0_resize_test 00:07:14.221 ************************************ 00:07:14.221 15:11:42 bdev_raid -- bdev/bdev_raid.sh@964 -- # run_test raid1_resize_test raid_resize_test 1 00:07:14.221 15:11:42 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:07:14.221 15:11:42 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:14.221 15:11:42 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:14.221 ************************************ 00:07:14.221 START TEST raid1_resize_test 00:07:14.221 ************************************ 00:07:14.221 15:11:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@1129 -- # raid_resize_test 1 00:07:14.221 15:11:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@332 -- # local raid_level=1 00:07:14.221 15:11:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@333 -- # local blksize=512 00:07:14.221 15:11:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@334 -- # local bdev_size_mb=32 00:07:14.221 15:11:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@335 -- # local new_bdev_size_mb=64 00:07:14.221 15:11:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@336 -- # local blkcnt 00:07:14.221 15:11:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@337 -- # local raid_size_mb 00:07:14.221 15:11:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@338 -- # local new_raid_size_mb 00:07:14.221 15:11:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@339 -- # local expected_size 00:07:14.221 15:11:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@342 -- # raid_pid=72228 00:07:14.221 Process raid pid: 72228 00:07:14.221 15:11:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@341 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:14.221 15:11:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@343 -- # echo 'Process raid pid: 72228' 00:07:14.221 15:11:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@344 -- # waitforlisten 72228 00:07:14.221 15:11:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@835 -- # '[' -z 72228 ']' 00:07:14.221 15:11:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:14.221 15:11:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:14.221 15:11:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:14.221 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:14.221 15:11:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:14.221 15:11:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:14.221 [2024-11-27 15:11:42.267248] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:07:14.221 [2024-11-27 15:11:42.267732] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:14.482 [2024-11-27 15:11:42.433535] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:14.482 [2024-11-27 15:11:42.459326] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:14.482 [2024-11-27 15:11:42.503487] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:14.482 [2024-11-27 15:11:42.503535] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:15.052 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:15.052 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@868 -- # return 0 00:07:15.052 15:11:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@346 -- # rpc_cmd bdev_null_create Base_1 32 512 00:07:15.052 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:15.052 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.052 Base_1 00:07:15.052 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:15.052 15:11:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@347 -- # rpc_cmd bdev_null_create Base_2 32 512 00:07:15.052 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:15.052 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.052 Base_2 00:07:15.052 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:15.052 15:11:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@349 -- # '[' 1 -eq 0 ']' 00:07:15.052 15:11:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@352 -- # rpc_cmd bdev_raid_create -r 1 -b ''\''Base_1 Base_2'\''' -n Raid 00:07:15.052 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:15.052 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.052 [2024-11-27 15:11:43.127106] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:07:15.052 [2024-11-27 15:11:43.128989] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:07:15.052 [2024-11-27 15:11:43.129048] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:07:15.052 [2024-11-27 15:11:43.129060] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:07:15.052 [2024-11-27 15:11:43.129309] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005a00 00:07:15.052 [2024-11-27 15:11:43.129438] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:07:15.052 [2024-11-27 15:11:43.129457] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000006280 00:07:15.052 [2024-11-27 15:11:43.129578] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:15.052 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:15.052 15:11:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@356 -- # rpc_cmd bdev_null_resize Base_1 64 00:07:15.052 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:15.052 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.052 [2024-11-27 15:11:43.139069] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:07:15.052 [2024-11-27 15:11:43.139097] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_1' was resized: old size 65536, new size 131072 00:07:15.052 true 00:07:15.052 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:15.052 15:11:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:15.052 15:11:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # jq '.[].num_blocks' 00:07:15.052 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:15.052 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.052 [2024-11-27 15:11:43.155211] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:15.312 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:15.312 15:11:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # blkcnt=65536 00:07:15.312 15:11:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@360 -- # raid_size_mb=32 00:07:15.312 15:11:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@361 -- # '[' 1 -eq 0 ']' 00:07:15.312 15:11:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@364 -- # expected_size=32 00:07:15.312 15:11:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@366 -- # '[' 32 '!=' 32 ']' 00:07:15.312 15:11:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@372 -- # rpc_cmd bdev_null_resize Base_2 64 00:07:15.312 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:15.312 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.312 [2024-11-27 15:11:43.199029] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:07:15.312 [2024-11-27 15:11:43.199054] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_2' was resized: old size 65536, new size 131072 00:07:15.312 [2024-11-27 15:11:43.199091] bdev_raid.c:2344:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 65536 to 131072 00:07:15.312 true 00:07:15.312 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:15.312 15:11:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:15.312 15:11:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # jq '.[].num_blocks' 00:07:15.312 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:15.312 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.312 [2024-11-27 15:11:43.215167] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:15.312 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:15.312 15:11:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # blkcnt=131072 00:07:15.312 15:11:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@376 -- # raid_size_mb=64 00:07:15.312 15:11:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@377 -- # '[' 1 -eq 0 ']' 00:07:15.312 15:11:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@380 -- # expected_size=64 00:07:15.312 15:11:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@382 -- # '[' 64 '!=' 64 ']' 00:07:15.312 15:11:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@387 -- # killprocess 72228 00:07:15.312 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@954 -- # '[' -z 72228 ']' 00:07:15.312 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@958 -- # kill -0 72228 00:07:15.312 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@959 -- # uname 00:07:15.312 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:15.312 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72228 00:07:15.312 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:15.312 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:15.312 killing process with pid 72228 00:07:15.312 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72228' 00:07:15.312 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@973 -- # kill 72228 00:07:15.312 [2024-11-27 15:11:43.297639] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:15.312 [2024-11-27 15:11:43.297733] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:15.312 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@978 -- # wait 72228 00:07:15.312 [2024-11-27 15:11:43.298178] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:15.312 [2024-11-27 15:11:43.298202] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Raid, state offline 00:07:15.312 [2024-11-27 15:11:43.299347] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:15.599 15:11:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@389 -- # return 0 00:07:15.599 00:07:15.599 real 0m1.334s 00:07:15.599 user 0m1.499s 00:07:15.599 sys 0m0.314s 00:07:15.599 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:15.599 15:11:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.599 ************************************ 00:07:15.599 END TEST raid1_resize_test 00:07:15.599 ************************************ 00:07:15.599 15:11:43 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:07:15.599 15:11:43 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:07:15.599 15:11:43 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 2 false 00:07:15.599 15:11:43 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:15.599 15:11:43 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:15.599 15:11:43 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:15.599 ************************************ 00:07:15.599 START TEST raid_state_function_test 00:07:15.599 ************************************ 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 2 false 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=72279 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 72279' 00:07:15.599 Process raid pid: 72279 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 72279 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 72279 ']' 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:15.599 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:15.599 15:11:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.599 [2024-11-27 15:11:43.675006] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:07:15.599 [2024-11-27 15:11:43.675148] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:15.859 [2024-11-27 15:11:43.847437] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:15.859 [2024-11-27 15:11:43.873627] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:15.859 [2024-11-27 15:11:43.915753] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:15.859 [2024-11-27 15:11:43.915799] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:16.430 15:11:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:16.430 15:11:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:07:16.430 15:11:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:16.430 15:11:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:16.430 15:11:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:16.430 [2024-11-27 15:11:44.510766] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:16.430 [2024-11-27 15:11:44.510822] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:16.430 [2024-11-27 15:11:44.510832] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:16.430 [2024-11-27 15:11:44.510841] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:16.430 15:11:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:16.430 15:11:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:16.430 15:11:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:16.430 15:11:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:16.430 15:11:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:16.430 15:11:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:16.430 15:11:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:16.430 15:11:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:16.430 15:11:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:16.430 15:11:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:16.430 15:11:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:16.430 15:11:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:16.430 15:11:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:16.430 15:11:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:16.430 15:11:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:16.690 15:11:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:16.690 15:11:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:16.690 "name": "Existed_Raid", 00:07:16.690 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:16.690 "strip_size_kb": 64, 00:07:16.690 "state": "configuring", 00:07:16.690 "raid_level": "raid0", 00:07:16.690 "superblock": false, 00:07:16.690 "num_base_bdevs": 2, 00:07:16.690 "num_base_bdevs_discovered": 0, 00:07:16.690 "num_base_bdevs_operational": 2, 00:07:16.690 "base_bdevs_list": [ 00:07:16.690 { 00:07:16.690 "name": "BaseBdev1", 00:07:16.690 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:16.690 "is_configured": false, 00:07:16.690 "data_offset": 0, 00:07:16.690 "data_size": 0 00:07:16.690 }, 00:07:16.690 { 00:07:16.690 "name": "BaseBdev2", 00:07:16.690 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:16.690 "is_configured": false, 00:07:16.691 "data_offset": 0, 00:07:16.691 "data_size": 0 00:07:16.691 } 00:07:16.691 ] 00:07:16.691 }' 00:07:16.691 15:11:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:16.691 15:11:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:16.951 [2024-11-27 15:11:44.930016] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:16.951 [2024-11-27 15:11:44.930063] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:16.951 [2024-11-27 15:11:44.941940] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:16.951 [2024-11-27 15:11:44.941974] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:16.951 [2024-11-27 15:11:44.941999] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:16.951 [2024-11-27 15:11:44.942008] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:16.951 [2024-11-27 15:11:44.962703] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:16.951 BaseBdev1 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:16.951 [ 00:07:16.951 { 00:07:16.951 "name": "BaseBdev1", 00:07:16.951 "aliases": [ 00:07:16.951 "7af47878-6a5f-4824-83c6-1073e9681771" 00:07:16.951 ], 00:07:16.951 "product_name": "Malloc disk", 00:07:16.951 "block_size": 512, 00:07:16.951 "num_blocks": 65536, 00:07:16.951 "uuid": "7af47878-6a5f-4824-83c6-1073e9681771", 00:07:16.951 "assigned_rate_limits": { 00:07:16.951 "rw_ios_per_sec": 0, 00:07:16.951 "rw_mbytes_per_sec": 0, 00:07:16.951 "r_mbytes_per_sec": 0, 00:07:16.951 "w_mbytes_per_sec": 0 00:07:16.951 }, 00:07:16.951 "claimed": true, 00:07:16.951 "claim_type": "exclusive_write", 00:07:16.951 "zoned": false, 00:07:16.951 "supported_io_types": { 00:07:16.951 "read": true, 00:07:16.951 "write": true, 00:07:16.951 "unmap": true, 00:07:16.951 "flush": true, 00:07:16.951 "reset": true, 00:07:16.951 "nvme_admin": false, 00:07:16.951 "nvme_io": false, 00:07:16.951 "nvme_io_md": false, 00:07:16.951 "write_zeroes": true, 00:07:16.951 "zcopy": true, 00:07:16.951 "get_zone_info": false, 00:07:16.951 "zone_management": false, 00:07:16.951 "zone_append": false, 00:07:16.951 "compare": false, 00:07:16.951 "compare_and_write": false, 00:07:16.951 "abort": true, 00:07:16.951 "seek_hole": false, 00:07:16.951 "seek_data": false, 00:07:16.951 "copy": true, 00:07:16.951 "nvme_iov_md": false 00:07:16.951 }, 00:07:16.951 "memory_domains": [ 00:07:16.951 { 00:07:16.951 "dma_device_id": "system", 00:07:16.951 "dma_device_type": 1 00:07:16.951 }, 00:07:16.951 { 00:07:16.951 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:16.951 "dma_device_type": 2 00:07:16.951 } 00:07:16.951 ], 00:07:16.951 "driver_specific": {} 00:07:16.951 } 00:07:16.951 ] 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:16.951 15:11:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:16.951 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:16.951 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:16.951 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:16.951 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:16.951 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:16.951 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:16.951 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:16.951 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:16.951 "name": "Existed_Raid", 00:07:16.951 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:16.951 "strip_size_kb": 64, 00:07:16.951 "state": "configuring", 00:07:16.951 "raid_level": "raid0", 00:07:16.951 "superblock": false, 00:07:16.951 "num_base_bdevs": 2, 00:07:16.951 "num_base_bdevs_discovered": 1, 00:07:16.951 "num_base_bdevs_operational": 2, 00:07:16.951 "base_bdevs_list": [ 00:07:16.951 { 00:07:16.951 "name": "BaseBdev1", 00:07:16.951 "uuid": "7af47878-6a5f-4824-83c6-1073e9681771", 00:07:16.951 "is_configured": true, 00:07:16.951 "data_offset": 0, 00:07:16.951 "data_size": 65536 00:07:16.951 }, 00:07:16.951 { 00:07:16.951 "name": "BaseBdev2", 00:07:16.951 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:16.951 "is_configured": false, 00:07:16.951 "data_offset": 0, 00:07:16.951 "data_size": 0 00:07:16.951 } 00:07:16.951 ] 00:07:16.952 }' 00:07:16.952 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:16.952 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.521 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:17.521 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:17.521 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.521 [2024-11-27 15:11:45.386019] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:17.521 [2024-11-27 15:11:45.386070] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:07:17.521 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:17.521 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:17.521 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:17.521 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.521 [2024-11-27 15:11:45.398036] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:17.521 [2024-11-27 15:11:45.399928] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:17.521 [2024-11-27 15:11:45.399975] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:17.521 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:17.521 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:17.521 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:17.521 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:17.521 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:17.521 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:17.521 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:17.521 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:17.521 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:17.521 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:17.521 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:17.521 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:17.521 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:17.521 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:17.521 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:17.521 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:17.521 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.521 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:17.521 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:17.521 "name": "Existed_Raid", 00:07:17.521 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:17.521 "strip_size_kb": 64, 00:07:17.521 "state": "configuring", 00:07:17.521 "raid_level": "raid0", 00:07:17.521 "superblock": false, 00:07:17.521 "num_base_bdevs": 2, 00:07:17.521 "num_base_bdevs_discovered": 1, 00:07:17.521 "num_base_bdevs_operational": 2, 00:07:17.521 "base_bdevs_list": [ 00:07:17.521 { 00:07:17.521 "name": "BaseBdev1", 00:07:17.521 "uuid": "7af47878-6a5f-4824-83c6-1073e9681771", 00:07:17.521 "is_configured": true, 00:07:17.521 "data_offset": 0, 00:07:17.521 "data_size": 65536 00:07:17.521 }, 00:07:17.521 { 00:07:17.521 "name": "BaseBdev2", 00:07:17.521 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:17.521 "is_configured": false, 00:07:17.521 "data_offset": 0, 00:07:17.521 "data_size": 0 00:07:17.521 } 00:07:17.521 ] 00:07:17.521 }' 00:07:17.521 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:17.521 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.781 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:17.781 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:17.781 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.781 [2024-11-27 15:11:45.864376] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:17.781 [2024-11-27 15:11:45.864424] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:07:17.781 [2024-11-27 15:11:45.864434] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:07:17.781 [2024-11-27 15:11:45.864732] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:07:17.781 [2024-11-27 15:11:45.864893] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:07:17.781 [2024-11-27 15:11:45.864928] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:07:17.781 [2024-11-27 15:11:45.865160] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:17.781 BaseBdev2 00:07:17.781 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:17.781 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:17.781 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:07:17.781 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:17.781 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:17.781 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:17.781 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:17.781 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:17.781 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:17.781 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.781 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:17.781 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:17.781 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:17.781 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.041 [ 00:07:18.041 { 00:07:18.041 "name": "BaseBdev2", 00:07:18.041 "aliases": [ 00:07:18.041 "e8251ba0-3524-45dd-9d2e-cfb0c19fd306" 00:07:18.041 ], 00:07:18.041 "product_name": "Malloc disk", 00:07:18.041 "block_size": 512, 00:07:18.041 "num_blocks": 65536, 00:07:18.041 "uuid": "e8251ba0-3524-45dd-9d2e-cfb0c19fd306", 00:07:18.041 "assigned_rate_limits": { 00:07:18.041 "rw_ios_per_sec": 0, 00:07:18.041 "rw_mbytes_per_sec": 0, 00:07:18.041 "r_mbytes_per_sec": 0, 00:07:18.041 "w_mbytes_per_sec": 0 00:07:18.041 }, 00:07:18.041 "claimed": true, 00:07:18.041 "claim_type": "exclusive_write", 00:07:18.041 "zoned": false, 00:07:18.041 "supported_io_types": { 00:07:18.041 "read": true, 00:07:18.041 "write": true, 00:07:18.041 "unmap": true, 00:07:18.041 "flush": true, 00:07:18.041 "reset": true, 00:07:18.041 "nvme_admin": false, 00:07:18.041 "nvme_io": false, 00:07:18.041 "nvme_io_md": false, 00:07:18.041 "write_zeroes": true, 00:07:18.041 "zcopy": true, 00:07:18.041 "get_zone_info": false, 00:07:18.041 "zone_management": false, 00:07:18.041 "zone_append": false, 00:07:18.041 "compare": false, 00:07:18.041 "compare_and_write": false, 00:07:18.041 "abort": true, 00:07:18.041 "seek_hole": false, 00:07:18.041 "seek_data": false, 00:07:18.041 "copy": true, 00:07:18.041 "nvme_iov_md": false 00:07:18.041 }, 00:07:18.041 "memory_domains": [ 00:07:18.041 { 00:07:18.041 "dma_device_id": "system", 00:07:18.041 "dma_device_type": 1 00:07:18.041 }, 00:07:18.041 { 00:07:18.041 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:18.041 "dma_device_type": 2 00:07:18.041 } 00:07:18.041 ], 00:07:18.041 "driver_specific": {} 00:07:18.041 } 00:07:18.041 ] 00:07:18.041 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:18.041 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:18.041 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:18.041 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:18.041 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 2 00:07:18.041 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:18.041 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:18.041 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:18.041 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:18.041 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:18.041 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:18.041 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:18.041 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:18.041 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:18.041 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:18.041 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:18.041 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:18.041 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.041 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:18.041 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:18.041 "name": "Existed_Raid", 00:07:18.041 "uuid": "516a11cd-8e3c-41d9-a1cf-f3a1185ff415", 00:07:18.041 "strip_size_kb": 64, 00:07:18.041 "state": "online", 00:07:18.041 "raid_level": "raid0", 00:07:18.041 "superblock": false, 00:07:18.041 "num_base_bdevs": 2, 00:07:18.041 "num_base_bdevs_discovered": 2, 00:07:18.041 "num_base_bdevs_operational": 2, 00:07:18.041 "base_bdevs_list": [ 00:07:18.041 { 00:07:18.041 "name": "BaseBdev1", 00:07:18.041 "uuid": "7af47878-6a5f-4824-83c6-1073e9681771", 00:07:18.041 "is_configured": true, 00:07:18.041 "data_offset": 0, 00:07:18.041 "data_size": 65536 00:07:18.041 }, 00:07:18.041 { 00:07:18.041 "name": "BaseBdev2", 00:07:18.041 "uuid": "e8251ba0-3524-45dd-9d2e-cfb0c19fd306", 00:07:18.041 "is_configured": true, 00:07:18.041 "data_offset": 0, 00:07:18.041 "data_size": 65536 00:07:18.041 } 00:07:18.041 ] 00:07:18.041 }' 00:07:18.041 15:11:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:18.041 15:11:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.302 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:18.302 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:18.302 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:18.302 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:18.302 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:18.302 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:18.302 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:18.302 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:18.302 15:11:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:18.302 15:11:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.302 [2024-11-27 15:11:46.367977] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:18.302 15:11:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:18.561 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:18.561 "name": "Existed_Raid", 00:07:18.561 "aliases": [ 00:07:18.561 "516a11cd-8e3c-41d9-a1cf-f3a1185ff415" 00:07:18.561 ], 00:07:18.561 "product_name": "Raid Volume", 00:07:18.561 "block_size": 512, 00:07:18.561 "num_blocks": 131072, 00:07:18.562 "uuid": "516a11cd-8e3c-41d9-a1cf-f3a1185ff415", 00:07:18.562 "assigned_rate_limits": { 00:07:18.562 "rw_ios_per_sec": 0, 00:07:18.562 "rw_mbytes_per_sec": 0, 00:07:18.562 "r_mbytes_per_sec": 0, 00:07:18.562 "w_mbytes_per_sec": 0 00:07:18.562 }, 00:07:18.562 "claimed": false, 00:07:18.562 "zoned": false, 00:07:18.562 "supported_io_types": { 00:07:18.562 "read": true, 00:07:18.562 "write": true, 00:07:18.562 "unmap": true, 00:07:18.562 "flush": true, 00:07:18.562 "reset": true, 00:07:18.562 "nvme_admin": false, 00:07:18.562 "nvme_io": false, 00:07:18.562 "nvme_io_md": false, 00:07:18.562 "write_zeroes": true, 00:07:18.562 "zcopy": false, 00:07:18.562 "get_zone_info": false, 00:07:18.562 "zone_management": false, 00:07:18.562 "zone_append": false, 00:07:18.562 "compare": false, 00:07:18.562 "compare_and_write": false, 00:07:18.562 "abort": false, 00:07:18.562 "seek_hole": false, 00:07:18.562 "seek_data": false, 00:07:18.562 "copy": false, 00:07:18.562 "nvme_iov_md": false 00:07:18.562 }, 00:07:18.562 "memory_domains": [ 00:07:18.562 { 00:07:18.562 "dma_device_id": "system", 00:07:18.562 "dma_device_type": 1 00:07:18.562 }, 00:07:18.562 { 00:07:18.562 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:18.562 "dma_device_type": 2 00:07:18.562 }, 00:07:18.562 { 00:07:18.562 "dma_device_id": "system", 00:07:18.562 "dma_device_type": 1 00:07:18.562 }, 00:07:18.562 { 00:07:18.562 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:18.562 "dma_device_type": 2 00:07:18.562 } 00:07:18.562 ], 00:07:18.562 "driver_specific": { 00:07:18.562 "raid": { 00:07:18.562 "uuid": "516a11cd-8e3c-41d9-a1cf-f3a1185ff415", 00:07:18.562 "strip_size_kb": 64, 00:07:18.562 "state": "online", 00:07:18.562 "raid_level": "raid0", 00:07:18.562 "superblock": false, 00:07:18.562 "num_base_bdevs": 2, 00:07:18.562 "num_base_bdevs_discovered": 2, 00:07:18.562 "num_base_bdevs_operational": 2, 00:07:18.562 "base_bdevs_list": [ 00:07:18.562 { 00:07:18.562 "name": "BaseBdev1", 00:07:18.562 "uuid": "7af47878-6a5f-4824-83c6-1073e9681771", 00:07:18.562 "is_configured": true, 00:07:18.562 "data_offset": 0, 00:07:18.562 "data_size": 65536 00:07:18.562 }, 00:07:18.562 { 00:07:18.562 "name": "BaseBdev2", 00:07:18.562 "uuid": "e8251ba0-3524-45dd-9d2e-cfb0c19fd306", 00:07:18.562 "is_configured": true, 00:07:18.562 "data_offset": 0, 00:07:18.562 "data_size": 65536 00:07:18.562 } 00:07:18.562 ] 00:07:18.562 } 00:07:18.562 } 00:07:18.562 }' 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:18.562 BaseBdev2' 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.562 [2024-11-27 15:11:46.591322] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:18.562 [2024-11-27 15:11:46.591359] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:18.562 [2024-11-27 15:11:46.591429] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 1 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:18.562 "name": "Existed_Raid", 00:07:18.562 "uuid": "516a11cd-8e3c-41d9-a1cf-f3a1185ff415", 00:07:18.562 "strip_size_kb": 64, 00:07:18.562 "state": "offline", 00:07:18.562 "raid_level": "raid0", 00:07:18.562 "superblock": false, 00:07:18.562 "num_base_bdevs": 2, 00:07:18.562 "num_base_bdevs_discovered": 1, 00:07:18.562 "num_base_bdevs_operational": 1, 00:07:18.562 "base_bdevs_list": [ 00:07:18.562 { 00:07:18.562 "name": null, 00:07:18.562 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:18.562 "is_configured": false, 00:07:18.562 "data_offset": 0, 00:07:18.562 "data_size": 65536 00:07:18.562 }, 00:07:18.562 { 00:07:18.562 "name": "BaseBdev2", 00:07:18.562 "uuid": "e8251ba0-3524-45dd-9d2e-cfb0c19fd306", 00:07:18.562 "is_configured": true, 00:07:18.562 "data_offset": 0, 00:07:18.562 "data_size": 65536 00:07:18.562 } 00:07:18.562 ] 00:07:18.562 }' 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:18.562 15:11:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.132 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:19.132 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:19.132 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:19.132 15:11:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:19.132 15:11:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:19.132 15:11:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.132 15:11:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:19.133 15:11:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:19.133 15:11:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:19.133 15:11:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:19.133 15:11:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:19.133 15:11:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.133 [2024-11-27 15:11:47.017997] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:19.133 [2024-11-27 15:11:47.018056] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:07:19.133 15:11:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:19.133 15:11:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:19.133 15:11:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:19.133 15:11:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:19.133 15:11:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:19.133 15:11:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:19.133 15:11:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.133 15:11:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:19.133 15:11:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:19.133 15:11:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:19.133 15:11:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:19.133 15:11:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 72279 00:07:19.133 15:11:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 72279 ']' 00:07:19.133 15:11:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 72279 00:07:19.133 15:11:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:07:19.133 15:11:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:19.133 15:11:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72279 00:07:19.133 15:11:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:19.133 15:11:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:19.133 killing process with pid 72279 00:07:19.133 15:11:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72279' 00:07:19.133 15:11:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 72279 00:07:19.133 [2024-11-27 15:11:47.126974] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:19.133 15:11:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 72279 00:07:19.133 [2024-11-27 15:11:47.128003] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:19.392 15:11:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:07:19.392 00:07:19.392 real 0m3.760s 00:07:19.392 user 0m5.942s 00:07:19.392 sys 0m0.751s 00:07:19.392 15:11:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:19.392 15:11:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.392 ************************************ 00:07:19.392 END TEST raid_state_function_test 00:07:19.392 ************************************ 00:07:19.392 15:11:47 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 2 true 00:07:19.392 15:11:47 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:19.393 15:11:47 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:19.393 15:11:47 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:19.393 ************************************ 00:07:19.393 START TEST raid_state_function_test_sb 00:07:19.393 ************************************ 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 2 true 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=72516 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:19.393 Process raid pid: 72516 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 72516' 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 72516 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 72516 ']' 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:19.393 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:19.393 15:11:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:19.652 [2024-11-27 15:11:47.501665] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:07:19.652 [2024-11-27 15:11:47.501784] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:19.652 [2024-11-27 15:11:47.651840] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:19.652 [2024-11-27 15:11:47.677235] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:19.652 [2024-11-27 15:11:47.720063] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:19.652 [2024-11-27 15:11:47.720104] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:20.222 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:20.222 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:07:20.222 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:20.222 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:20.222 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:20.222 [2024-11-27 15:11:48.324025] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:20.222 [2024-11-27 15:11:48.324075] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:20.222 [2024-11-27 15:11:48.324092] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:20.222 [2024-11-27 15:11:48.324103] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:20.481 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:20.481 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:20.481 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:20.481 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:20.481 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:20.481 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:20.481 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:20.481 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:20.481 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:20.481 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:20.481 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:20.481 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:20.481 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:20.481 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:20.481 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:20.481 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:20.481 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:20.481 "name": "Existed_Raid", 00:07:20.481 "uuid": "c0cf73d9-a52e-4ff0-97b9-0ae76764bd11", 00:07:20.481 "strip_size_kb": 64, 00:07:20.481 "state": "configuring", 00:07:20.481 "raid_level": "raid0", 00:07:20.481 "superblock": true, 00:07:20.481 "num_base_bdevs": 2, 00:07:20.481 "num_base_bdevs_discovered": 0, 00:07:20.481 "num_base_bdevs_operational": 2, 00:07:20.481 "base_bdevs_list": [ 00:07:20.481 { 00:07:20.481 "name": "BaseBdev1", 00:07:20.481 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:20.482 "is_configured": false, 00:07:20.482 "data_offset": 0, 00:07:20.482 "data_size": 0 00:07:20.482 }, 00:07:20.482 { 00:07:20.482 "name": "BaseBdev2", 00:07:20.482 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:20.482 "is_configured": false, 00:07:20.482 "data_offset": 0, 00:07:20.482 "data_size": 0 00:07:20.482 } 00:07:20.482 ] 00:07:20.482 }' 00:07:20.482 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:20.482 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:20.741 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:20.741 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:20.741 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:20.741 [2024-11-27 15:11:48.775236] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:20.741 [2024-11-27 15:11:48.775285] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:07:20.741 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:20.741 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:20.741 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:20.741 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:20.741 [2024-11-27 15:11:48.787213] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:20.741 [2024-11-27 15:11:48.787252] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:20.741 [2024-11-27 15:11:48.787261] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:20.741 [2024-11-27 15:11:48.787270] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:20.741 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:20.741 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:20.741 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:20.741 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:20.741 [2024-11-27 15:11:48.808102] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:20.741 BaseBdev1 00:07:20.741 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:20.741 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:20.741 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:07:20.741 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:20.741 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:20.741 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:20.741 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:20.741 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:20.741 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:20.741 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:20.741 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:20.741 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:20.742 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:20.742 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:20.742 [ 00:07:20.742 { 00:07:20.742 "name": "BaseBdev1", 00:07:20.742 "aliases": [ 00:07:20.742 "20a85064-1d48-4905-b9aa-ddc9a8faa4a1" 00:07:20.742 ], 00:07:20.742 "product_name": "Malloc disk", 00:07:20.742 "block_size": 512, 00:07:20.742 "num_blocks": 65536, 00:07:20.742 "uuid": "20a85064-1d48-4905-b9aa-ddc9a8faa4a1", 00:07:20.742 "assigned_rate_limits": { 00:07:20.742 "rw_ios_per_sec": 0, 00:07:20.742 "rw_mbytes_per_sec": 0, 00:07:20.742 "r_mbytes_per_sec": 0, 00:07:20.742 "w_mbytes_per_sec": 0 00:07:20.742 }, 00:07:20.742 "claimed": true, 00:07:20.742 "claim_type": "exclusive_write", 00:07:20.742 "zoned": false, 00:07:20.742 "supported_io_types": { 00:07:20.742 "read": true, 00:07:20.742 "write": true, 00:07:20.742 "unmap": true, 00:07:20.742 "flush": true, 00:07:20.742 "reset": true, 00:07:20.742 "nvme_admin": false, 00:07:20.742 "nvme_io": false, 00:07:20.742 "nvme_io_md": false, 00:07:20.742 "write_zeroes": true, 00:07:20.742 "zcopy": true, 00:07:20.742 "get_zone_info": false, 00:07:20.742 "zone_management": false, 00:07:20.742 "zone_append": false, 00:07:20.742 "compare": false, 00:07:20.742 "compare_and_write": false, 00:07:20.742 "abort": true, 00:07:20.742 "seek_hole": false, 00:07:20.742 "seek_data": false, 00:07:20.742 "copy": true, 00:07:20.742 "nvme_iov_md": false 00:07:20.742 }, 00:07:20.742 "memory_domains": [ 00:07:20.742 { 00:07:20.742 "dma_device_id": "system", 00:07:20.742 "dma_device_type": 1 00:07:20.742 }, 00:07:20.742 { 00:07:20.742 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:20.742 "dma_device_type": 2 00:07:20.742 } 00:07:20.742 ], 00:07:20.742 "driver_specific": {} 00:07:20.742 } 00:07:20.742 ] 00:07:20.742 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:20.742 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:20.742 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:20.742 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:20.742 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:21.001 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:21.001 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:21.001 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:21.001 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:21.001 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:21.001 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:21.001 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:21.001 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:21.001 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:21.001 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:21.001 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:21.001 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:21.001 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:21.001 "name": "Existed_Raid", 00:07:21.001 "uuid": "2f902a3b-912c-42e4-9145-b7aac8203e28", 00:07:21.001 "strip_size_kb": 64, 00:07:21.001 "state": "configuring", 00:07:21.001 "raid_level": "raid0", 00:07:21.001 "superblock": true, 00:07:21.001 "num_base_bdevs": 2, 00:07:21.001 "num_base_bdevs_discovered": 1, 00:07:21.001 "num_base_bdevs_operational": 2, 00:07:21.001 "base_bdevs_list": [ 00:07:21.001 { 00:07:21.001 "name": "BaseBdev1", 00:07:21.001 "uuid": "20a85064-1d48-4905-b9aa-ddc9a8faa4a1", 00:07:21.001 "is_configured": true, 00:07:21.001 "data_offset": 2048, 00:07:21.001 "data_size": 63488 00:07:21.001 }, 00:07:21.001 { 00:07:21.001 "name": "BaseBdev2", 00:07:21.001 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:21.001 "is_configured": false, 00:07:21.001 "data_offset": 0, 00:07:21.001 "data_size": 0 00:07:21.001 } 00:07:21.001 ] 00:07:21.001 }' 00:07:21.001 15:11:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:21.001 15:11:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:21.269 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:21.269 15:11:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:21.269 15:11:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:21.269 [2024-11-27 15:11:49.251597] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:21.269 [2024-11-27 15:11:49.251652] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:07:21.269 15:11:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:21.269 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:21.269 15:11:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:21.269 15:11:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:21.269 [2024-11-27 15:11:49.263610] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:21.269 [2024-11-27 15:11:49.265575] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:21.269 [2024-11-27 15:11:49.265614] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:21.269 15:11:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:21.269 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:21.269 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:21.269 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:21.269 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:21.269 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:21.269 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:21.269 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:21.270 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:21.270 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:21.270 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:21.270 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:21.270 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:21.270 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:21.270 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:21.270 15:11:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:21.270 15:11:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:21.270 15:11:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:21.270 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:21.270 "name": "Existed_Raid", 00:07:21.270 "uuid": "dbd508c5-3045-4d21-b329-0e8faa9ed405", 00:07:21.270 "strip_size_kb": 64, 00:07:21.270 "state": "configuring", 00:07:21.270 "raid_level": "raid0", 00:07:21.270 "superblock": true, 00:07:21.270 "num_base_bdevs": 2, 00:07:21.270 "num_base_bdevs_discovered": 1, 00:07:21.270 "num_base_bdevs_operational": 2, 00:07:21.270 "base_bdevs_list": [ 00:07:21.270 { 00:07:21.270 "name": "BaseBdev1", 00:07:21.270 "uuid": "20a85064-1d48-4905-b9aa-ddc9a8faa4a1", 00:07:21.270 "is_configured": true, 00:07:21.270 "data_offset": 2048, 00:07:21.270 "data_size": 63488 00:07:21.270 }, 00:07:21.270 { 00:07:21.270 "name": "BaseBdev2", 00:07:21.270 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:21.270 "is_configured": false, 00:07:21.270 "data_offset": 0, 00:07:21.270 "data_size": 0 00:07:21.270 } 00:07:21.270 ] 00:07:21.270 }' 00:07:21.270 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:21.270 15:11:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:21.839 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:21.839 15:11:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:21.839 15:11:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:21.839 [2024-11-27 15:11:49.693866] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:21.839 [2024-11-27 15:11:49.694087] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:07:21.840 [2024-11-27 15:11:49.694108] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:21.840 [2024-11-27 15:11:49.694416] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:07:21.840 [2024-11-27 15:11:49.694569] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:07:21.840 [2024-11-27 15:11:49.694598] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:07:21.840 BaseBdev2 00:07:21.840 [2024-11-27 15:11:49.694713] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:21.840 [ 00:07:21.840 { 00:07:21.840 "name": "BaseBdev2", 00:07:21.840 "aliases": [ 00:07:21.840 "4ed22db6-19bb-4272-80b5-db6599d54dc8" 00:07:21.840 ], 00:07:21.840 "product_name": "Malloc disk", 00:07:21.840 "block_size": 512, 00:07:21.840 "num_blocks": 65536, 00:07:21.840 "uuid": "4ed22db6-19bb-4272-80b5-db6599d54dc8", 00:07:21.840 "assigned_rate_limits": { 00:07:21.840 "rw_ios_per_sec": 0, 00:07:21.840 "rw_mbytes_per_sec": 0, 00:07:21.840 "r_mbytes_per_sec": 0, 00:07:21.840 "w_mbytes_per_sec": 0 00:07:21.840 }, 00:07:21.840 "claimed": true, 00:07:21.840 "claim_type": "exclusive_write", 00:07:21.840 "zoned": false, 00:07:21.840 "supported_io_types": { 00:07:21.840 "read": true, 00:07:21.840 "write": true, 00:07:21.840 "unmap": true, 00:07:21.840 "flush": true, 00:07:21.840 "reset": true, 00:07:21.840 "nvme_admin": false, 00:07:21.840 "nvme_io": false, 00:07:21.840 "nvme_io_md": false, 00:07:21.840 "write_zeroes": true, 00:07:21.840 "zcopy": true, 00:07:21.840 "get_zone_info": false, 00:07:21.840 "zone_management": false, 00:07:21.840 "zone_append": false, 00:07:21.840 "compare": false, 00:07:21.840 "compare_and_write": false, 00:07:21.840 "abort": true, 00:07:21.840 "seek_hole": false, 00:07:21.840 "seek_data": false, 00:07:21.840 "copy": true, 00:07:21.840 "nvme_iov_md": false 00:07:21.840 }, 00:07:21.840 "memory_domains": [ 00:07:21.840 { 00:07:21.840 "dma_device_id": "system", 00:07:21.840 "dma_device_type": 1 00:07:21.840 }, 00:07:21.840 { 00:07:21.840 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:21.840 "dma_device_type": 2 00:07:21.840 } 00:07:21.840 ], 00:07:21.840 "driver_specific": {} 00:07:21.840 } 00:07:21.840 ] 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 2 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:21.840 "name": "Existed_Raid", 00:07:21.840 "uuid": "dbd508c5-3045-4d21-b329-0e8faa9ed405", 00:07:21.840 "strip_size_kb": 64, 00:07:21.840 "state": "online", 00:07:21.840 "raid_level": "raid0", 00:07:21.840 "superblock": true, 00:07:21.840 "num_base_bdevs": 2, 00:07:21.840 "num_base_bdevs_discovered": 2, 00:07:21.840 "num_base_bdevs_operational": 2, 00:07:21.840 "base_bdevs_list": [ 00:07:21.840 { 00:07:21.840 "name": "BaseBdev1", 00:07:21.840 "uuid": "20a85064-1d48-4905-b9aa-ddc9a8faa4a1", 00:07:21.840 "is_configured": true, 00:07:21.840 "data_offset": 2048, 00:07:21.840 "data_size": 63488 00:07:21.840 }, 00:07:21.840 { 00:07:21.840 "name": "BaseBdev2", 00:07:21.840 "uuid": "4ed22db6-19bb-4272-80b5-db6599d54dc8", 00:07:21.840 "is_configured": true, 00:07:21.840 "data_offset": 2048, 00:07:21.840 "data_size": 63488 00:07:21.840 } 00:07:21.840 ] 00:07:21.840 }' 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:21.840 15:11:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.410 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:22.410 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:22.410 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:22.410 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:22.410 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:07:22.410 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:22.410 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:22.410 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:22.410 15:11:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.410 15:11:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.410 [2024-11-27 15:11:50.225324] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:22.410 15:11:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.410 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:22.410 "name": "Existed_Raid", 00:07:22.410 "aliases": [ 00:07:22.410 "dbd508c5-3045-4d21-b329-0e8faa9ed405" 00:07:22.410 ], 00:07:22.410 "product_name": "Raid Volume", 00:07:22.410 "block_size": 512, 00:07:22.410 "num_blocks": 126976, 00:07:22.410 "uuid": "dbd508c5-3045-4d21-b329-0e8faa9ed405", 00:07:22.410 "assigned_rate_limits": { 00:07:22.410 "rw_ios_per_sec": 0, 00:07:22.410 "rw_mbytes_per_sec": 0, 00:07:22.410 "r_mbytes_per_sec": 0, 00:07:22.410 "w_mbytes_per_sec": 0 00:07:22.410 }, 00:07:22.410 "claimed": false, 00:07:22.410 "zoned": false, 00:07:22.410 "supported_io_types": { 00:07:22.410 "read": true, 00:07:22.410 "write": true, 00:07:22.410 "unmap": true, 00:07:22.410 "flush": true, 00:07:22.410 "reset": true, 00:07:22.410 "nvme_admin": false, 00:07:22.410 "nvme_io": false, 00:07:22.410 "nvme_io_md": false, 00:07:22.410 "write_zeroes": true, 00:07:22.410 "zcopy": false, 00:07:22.410 "get_zone_info": false, 00:07:22.410 "zone_management": false, 00:07:22.410 "zone_append": false, 00:07:22.410 "compare": false, 00:07:22.410 "compare_and_write": false, 00:07:22.410 "abort": false, 00:07:22.410 "seek_hole": false, 00:07:22.410 "seek_data": false, 00:07:22.410 "copy": false, 00:07:22.410 "nvme_iov_md": false 00:07:22.410 }, 00:07:22.410 "memory_domains": [ 00:07:22.410 { 00:07:22.410 "dma_device_id": "system", 00:07:22.410 "dma_device_type": 1 00:07:22.410 }, 00:07:22.410 { 00:07:22.410 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:22.410 "dma_device_type": 2 00:07:22.410 }, 00:07:22.410 { 00:07:22.410 "dma_device_id": "system", 00:07:22.410 "dma_device_type": 1 00:07:22.410 }, 00:07:22.410 { 00:07:22.410 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:22.410 "dma_device_type": 2 00:07:22.410 } 00:07:22.410 ], 00:07:22.410 "driver_specific": { 00:07:22.410 "raid": { 00:07:22.410 "uuid": "dbd508c5-3045-4d21-b329-0e8faa9ed405", 00:07:22.410 "strip_size_kb": 64, 00:07:22.410 "state": "online", 00:07:22.410 "raid_level": "raid0", 00:07:22.410 "superblock": true, 00:07:22.410 "num_base_bdevs": 2, 00:07:22.410 "num_base_bdevs_discovered": 2, 00:07:22.410 "num_base_bdevs_operational": 2, 00:07:22.410 "base_bdevs_list": [ 00:07:22.410 { 00:07:22.410 "name": "BaseBdev1", 00:07:22.410 "uuid": "20a85064-1d48-4905-b9aa-ddc9a8faa4a1", 00:07:22.410 "is_configured": true, 00:07:22.410 "data_offset": 2048, 00:07:22.410 "data_size": 63488 00:07:22.410 }, 00:07:22.410 { 00:07:22.410 "name": "BaseBdev2", 00:07:22.410 "uuid": "4ed22db6-19bb-4272-80b5-db6599d54dc8", 00:07:22.410 "is_configured": true, 00:07:22.410 "data_offset": 2048, 00:07:22.410 "data_size": 63488 00:07:22.410 } 00:07:22.410 ] 00:07:22.410 } 00:07:22.410 } 00:07:22.410 }' 00:07:22.410 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:22.410 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:22.410 BaseBdev2' 00:07:22.410 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:22.410 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:22.410 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.411 [2024-11-27 15:11:50.448756] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:22.411 [2024-11-27 15:11:50.448790] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:22.411 [2024-11-27 15:11:50.448852] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 1 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:22.411 15:11:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.671 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:22.671 "name": "Existed_Raid", 00:07:22.671 "uuid": "dbd508c5-3045-4d21-b329-0e8faa9ed405", 00:07:22.671 "strip_size_kb": 64, 00:07:22.671 "state": "offline", 00:07:22.671 "raid_level": "raid0", 00:07:22.671 "superblock": true, 00:07:22.671 "num_base_bdevs": 2, 00:07:22.671 "num_base_bdevs_discovered": 1, 00:07:22.671 "num_base_bdevs_operational": 1, 00:07:22.671 "base_bdevs_list": [ 00:07:22.671 { 00:07:22.671 "name": null, 00:07:22.671 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:22.671 "is_configured": false, 00:07:22.671 "data_offset": 0, 00:07:22.671 "data_size": 63488 00:07:22.671 }, 00:07:22.671 { 00:07:22.671 "name": "BaseBdev2", 00:07:22.671 "uuid": "4ed22db6-19bb-4272-80b5-db6599d54dc8", 00:07:22.671 "is_configured": true, 00:07:22.671 "data_offset": 2048, 00:07:22.671 "data_size": 63488 00:07:22.671 } 00:07:22.671 ] 00:07:22.671 }' 00:07:22.671 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:22.671 15:11:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.930 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:22.930 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:22.930 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:22.930 15:11:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.930 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:22.930 15:11:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.930 15:11:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.930 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:22.930 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:22.930 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:22.930 15:11:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.930 15:11:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.930 [2024-11-27 15:11:50.951319] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:22.930 [2024-11-27 15:11:50.951378] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:07:22.930 15:11:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.930 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:22.930 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:22.930 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:22.930 15:11:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:22.930 15:11:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.930 15:11:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.930 15:11:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.930 15:11:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:22.930 15:11:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:22.930 15:11:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:22.930 15:11:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 72516 00:07:22.930 15:11:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 72516 ']' 00:07:22.930 15:11:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 72516 00:07:22.930 15:11:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:07:22.930 15:11:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:22.930 15:11:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72516 00:07:23.189 15:11:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:23.189 15:11:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:23.189 killing process with pid 72516 00:07:23.189 15:11:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72516' 00:07:23.189 15:11:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 72516 00:07:23.189 [2024-11-27 15:11:51.044816] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:23.189 15:11:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 72516 00:07:23.189 [2024-11-27 15:11:51.045866] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:23.189 15:11:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:07:23.189 00:07:23.189 real 0m3.856s 00:07:23.189 user 0m6.109s 00:07:23.189 sys 0m0.757s 00:07:23.190 15:11:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:23.190 15:11:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:23.190 ************************************ 00:07:23.190 END TEST raid_state_function_test_sb 00:07:23.190 ************************************ 00:07:23.448 15:11:51 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 2 00:07:23.448 15:11:51 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:07:23.448 15:11:51 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:23.448 15:11:51 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:23.448 ************************************ 00:07:23.448 START TEST raid_superblock_test 00:07:23.448 ************************************ 00:07:23.448 15:11:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid0 2 00:07:23.448 15:11:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:07:23.448 15:11:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:07:23.448 15:11:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:07:23.448 15:11:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:07:23.448 15:11:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:07:23.448 15:11:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:07:23.448 15:11:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:07:23.448 15:11:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:07:23.448 15:11:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:07:23.448 15:11:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:07:23.448 15:11:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:07:23.448 15:11:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:07:23.448 15:11:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:07:23.448 15:11:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:07:23.448 15:11:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:07:23.448 15:11:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:07:23.448 15:11:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=72757 00:07:23.448 15:11:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 72757 00:07:23.448 15:11:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:07:23.448 15:11:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 72757 ']' 00:07:23.448 15:11:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:23.448 15:11:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:23.448 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:23.448 15:11:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:23.448 15:11:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:23.448 15:11:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:23.448 [2024-11-27 15:11:51.429152] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:07:23.448 [2024-11-27 15:11:51.429297] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72757 ] 00:07:23.707 [2024-11-27 15:11:51.605050] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:23.707 [2024-11-27 15:11:51.631370] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:23.707 [2024-11-27 15:11:51.674064] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:23.707 [2024-11-27 15:11:51.674107] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.283 malloc1 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.283 [2024-11-27 15:11:52.266025] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:24.283 [2024-11-27 15:11:52.266088] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:24.283 [2024-11-27 15:11:52.266115] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:07:24.283 [2024-11-27 15:11:52.266131] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:24.283 [2024-11-27 15:11:52.268257] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:24.283 [2024-11-27 15:11:52.268294] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:24.283 pt1 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.283 malloc2 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:24.283 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.283 [2024-11-27 15:11:52.294516] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:24.284 [2024-11-27 15:11:52.294564] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:24.284 [2024-11-27 15:11:52.294580] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:07:24.284 [2024-11-27 15:11:52.294589] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:24.284 [2024-11-27 15:11:52.296691] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:24.284 [2024-11-27 15:11:52.296725] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:24.284 pt2 00:07:24.284 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:24.284 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:24.284 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:24.284 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:07:24.284 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:24.284 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.284 [2024-11-27 15:11:52.306534] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:24.284 [2024-11-27 15:11:52.308393] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:24.284 [2024-11-27 15:11:52.308531] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:07:24.284 [2024-11-27 15:11:52.308545] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:24.284 [2024-11-27 15:11:52.308783] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:07:24.284 [2024-11-27 15:11:52.308925] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:07:24.284 [2024-11-27 15:11:52.308939] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:07:24.284 [2024-11-27 15:11:52.309065] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:24.284 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:24.284 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:24.284 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:24.284 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:24.284 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:24.284 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:24.284 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:24.284 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:24.284 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:24.284 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:24.284 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:24.284 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:24.284 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:24.284 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:24.284 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.284 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:24.284 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:24.284 "name": "raid_bdev1", 00:07:24.284 "uuid": "921efa90-8be7-4233-93dc-1f685a9600e7", 00:07:24.284 "strip_size_kb": 64, 00:07:24.284 "state": "online", 00:07:24.284 "raid_level": "raid0", 00:07:24.284 "superblock": true, 00:07:24.284 "num_base_bdevs": 2, 00:07:24.284 "num_base_bdevs_discovered": 2, 00:07:24.284 "num_base_bdevs_operational": 2, 00:07:24.284 "base_bdevs_list": [ 00:07:24.284 { 00:07:24.284 "name": "pt1", 00:07:24.284 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:24.284 "is_configured": true, 00:07:24.284 "data_offset": 2048, 00:07:24.284 "data_size": 63488 00:07:24.284 }, 00:07:24.284 { 00:07:24.284 "name": "pt2", 00:07:24.284 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:24.284 "is_configured": true, 00:07:24.284 "data_offset": 2048, 00:07:24.284 "data_size": 63488 00:07:24.284 } 00:07:24.284 ] 00:07:24.284 }' 00:07:24.284 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:24.284 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.859 [2024-11-27 15:11:52.722202] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:24.859 "name": "raid_bdev1", 00:07:24.859 "aliases": [ 00:07:24.859 "921efa90-8be7-4233-93dc-1f685a9600e7" 00:07:24.859 ], 00:07:24.859 "product_name": "Raid Volume", 00:07:24.859 "block_size": 512, 00:07:24.859 "num_blocks": 126976, 00:07:24.859 "uuid": "921efa90-8be7-4233-93dc-1f685a9600e7", 00:07:24.859 "assigned_rate_limits": { 00:07:24.859 "rw_ios_per_sec": 0, 00:07:24.859 "rw_mbytes_per_sec": 0, 00:07:24.859 "r_mbytes_per_sec": 0, 00:07:24.859 "w_mbytes_per_sec": 0 00:07:24.859 }, 00:07:24.859 "claimed": false, 00:07:24.859 "zoned": false, 00:07:24.859 "supported_io_types": { 00:07:24.859 "read": true, 00:07:24.859 "write": true, 00:07:24.859 "unmap": true, 00:07:24.859 "flush": true, 00:07:24.859 "reset": true, 00:07:24.859 "nvme_admin": false, 00:07:24.859 "nvme_io": false, 00:07:24.859 "nvme_io_md": false, 00:07:24.859 "write_zeroes": true, 00:07:24.859 "zcopy": false, 00:07:24.859 "get_zone_info": false, 00:07:24.859 "zone_management": false, 00:07:24.859 "zone_append": false, 00:07:24.859 "compare": false, 00:07:24.859 "compare_and_write": false, 00:07:24.859 "abort": false, 00:07:24.859 "seek_hole": false, 00:07:24.859 "seek_data": false, 00:07:24.859 "copy": false, 00:07:24.859 "nvme_iov_md": false 00:07:24.859 }, 00:07:24.859 "memory_domains": [ 00:07:24.859 { 00:07:24.859 "dma_device_id": "system", 00:07:24.859 "dma_device_type": 1 00:07:24.859 }, 00:07:24.859 { 00:07:24.859 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:24.859 "dma_device_type": 2 00:07:24.859 }, 00:07:24.859 { 00:07:24.859 "dma_device_id": "system", 00:07:24.859 "dma_device_type": 1 00:07:24.859 }, 00:07:24.859 { 00:07:24.859 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:24.859 "dma_device_type": 2 00:07:24.859 } 00:07:24.859 ], 00:07:24.859 "driver_specific": { 00:07:24.859 "raid": { 00:07:24.859 "uuid": "921efa90-8be7-4233-93dc-1f685a9600e7", 00:07:24.859 "strip_size_kb": 64, 00:07:24.859 "state": "online", 00:07:24.859 "raid_level": "raid0", 00:07:24.859 "superblock": true, 00:07:24.859 "num_base_bdevs": 2, 00:07:24.859 "num_base_bdevs_discovered": 2, 00:07:24.859 "num_base_bdevs_operational": 2, 00:07:24.859 "base_bdevs_list": [ 00:07:24.859 { 00:07:24.859 "name": "pt1", 00:07:24.859 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:24.859 "is_configured": true, 00:07:24.859 "data_offset": 2048, 00:07:24.859 "data_size": 63488 00:07:24.859 }, 00:07:24.859 { 00:07:24.859 "name": "pt2", 00:07:24.859 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:24.859 "is_configured": true, 00:07:24.859 "data_offset": 2048, 00:07:24.859 "data_size": 63488 00:07:24.859 } 00:07:24.859 ] 00:07:24.859 } 00:07:24.859 } 00:07:24.859 }' 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:24.859 pt2' 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:24.859 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.859 [2024-11-27 15:11:52.945785] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:25.120 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:25.120 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=921efa90-8be7-4233-93dc-1f685a9600e7 00:07:25.120 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 921efa90-8be7-4233-93dc-1f685a9600e7 ']' 00:07:25.120 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:25.120 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:25.120 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.120 [2024-11-27 15:11:52.989386] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:25.120 [2024-11-27 15:11:52.989420] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:25.120 [2024-11-27 15:11:52.989493] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:25.120 [2024-11-27 15:11:52.989542] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:25.120 [2024-11-27 15:11:52.989560] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:07:25.120 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:25.120 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:25.120 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:25.120 15:11:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:07:25.120 15:11:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.120 [2024-11-27 15:11:53.129169] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:07:25.120 [2024-11-27 15:11:53.131073] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:07:25.120 [2024-11-27 15:11:53.131142] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:07:25.120 [2024-11-27 15:11:53.131178] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:07:25.120 [2024-11-27 15:11:53.131193] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:25.120 [2024-11-27 15:11:53.131201] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state configuring 00:07:25.120 request: 00:07:25.120 { 00:07:25.120 "name": "raid_bdev1", 00:07:25.120 "raid_level": "raid0", 00:07:25.120 "base_bdevs": [ 00:07:25.120 "malloc1", 00:07:25.120 "malloc2" 00:07:25.120 ], 00:07:25.120 "strip_size_kb": 64, 00:07:25.120 "superblock": false, 00:07:25.120 "method": "bdev_raid_create", 00:07:25.120 "req_id": 1 00:07:25.120 } 00:07:25.120 Got JSON-RPC error response 00:07:25.120 response: 00:07:25.120 { 00:07:25.120 "code": -17, 00:07:25.120 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:07:25.120 } 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.120 [2024-11-27 15:11:53.185028] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:25.120 [2024-11-27 15:11:53.185076] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:25.120 [2024-11-27 15:11:53.185096] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:07:25.120 [2024-11-27 15:11:53.185104] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:25.120 [2024-11-27 15:11:53.187232] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:25.120 [2024-11-27 15:11:53.187264] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:25.120 [2024-11-27 15:11:53.187334] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:07:25.120 [2024-11-27 15:11:53.187371] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:25.120 pt1 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 2 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:25.120 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:25.121 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:25.121 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:25.121 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:25.121 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:25.121 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:25.121 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:25.121 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.121 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:25.380 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:25.380 "name": "raid_bdev1", 00:07:25.380 "uuid": "921efa90-8be7-4233-93dc-1f685a9600e7", 00:07:25.380 "strip_size_kb": 64, 00:07:25.380 "state": "configuring", 00:07:25.380 "raid_level": "raid0", 00:07:25.380 "superblock": true, 00:07:25.380 "num_base_bdevs": 2, 00:07:25.380 "num_base_bdevs_discovered": 1, 00:07:25.380 "num_base_bdevs_operational": 2, 00:07:25.380 "base_bdevs_list": [ 00:07:25.380 { 00:07:25.380 "name": "pt1", 00:07:25.380 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:25.380 "is_configured": true, 00:07:25.380 "data_offset": 2048, 00:07:25.380 "data_size": 63488 00:07:25.380 }, 00:07:25.380 { 00:07:25.380 "name": null, 00:07:25.380 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:25.380 "is_configured": false, 00:07:25.380 "data_offset": 2048, 00:07:25.380 "data_size": 63488 00:07:25.380 } 00:07:25.380 ] 00:07:25.380 }' 00:07:25.380 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:25.380 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.640 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:07:25.640 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:07:25.640 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:25.640 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:25.640 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:25.640 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.640 [2024-11-27 15:11:53.612340] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:25.640 [2024-11-27 15:11:53.612405] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:25.640 [2024-11-27 15:11:53.612429] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:07:25.640 [2024-11-27 15:11:53.612439] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:25.640 [2024-11-27 15:11:53.612859] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:25.640 [2024-11-27 15:11:53.612876] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:25.640 [2024-11-27 15:11:53.612966] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:07:25.640 [2024-11-27 15:11:53.613015] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:25.640 [2024-11-27 15:11:53.613103] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:07:25.640 [2024-11-27 15:11:53.613111] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:25.640 [2024-11-27 15:11:53.613347] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:07:25.640 [2024-11-27 15:11:53.613456] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:07:25.640 [2024-11-27 15:11:53.613468] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:07:25.640 [2024-11-27 15:11:53.613564] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:25.640 pt2 00:07:25.640 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:25.640 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:07:25.640 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:25.640 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:25.640 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:25.640 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:25.640 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:25.640 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:25.640 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:25.640 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:25.640 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:25.640 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:25.640 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:25.640 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:25.640 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:25.640 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:25.640 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.640 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:25.640 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:25.640 "name": "raid_bdev1", 00:07:25.640 "uuid": "921efa90-8be7-4233-93dc-1f685a9600e7", 00:07:25.640 "strip_size_kb": 64, 00:07:25.640 "state": "online", 00:07:25.640 "raid_level": "raid0", 00:07:25.640 "superblock": true, 00:07:25.640 "num_base_bdevs": 2, 00:07:25.640 "num_base_bdevs_discovered": 2, 00:07:25.640 "num_base_bdevs_operational": 2, 00:07:25.640 "base_bdevs_list": [ 00:07:25.640 { 00:07:25.640 "name": "pt1", 00:07:25.640 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:25.640 "is_configured": true, 00:07:25.640 "data_offset": 2048, 00:07:25.640 "data_size": 63488 00:07:25.640 }, 00:07:25.640 { 00:07:25.640 "name": "pt2", 00:07:25.640 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:25.640 "is_configured": true, 00:07:25.640 "data_offset": 2048, 00:07:25.640 "data_size": 63488 00:07:25.640 } 00:07:25.640 ] 00:07:25.640 }' 00:07:25.640 15:11:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:25.640 15:11:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.210 15:11:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:07:26.210 15:11:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:26.210 15:11:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:26.210 15:11:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:26.210 15:11:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:26.210 15:11:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:26.210 15:11:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:26.210 15:11:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:26.210 15:11:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:26.210 15:11:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.210 [2024-11-27 15:11:54.067887] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:26.210 15:11:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:26.210 15:11:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:26.210 "name": "raid_bdev1", 00:07:26.210 "aliases": [ 00:07:26.210 "921efa90-8be7-4233-93dc-1f685a9600e7" 00:07:26.210 ], 00:07:26.210 "product_name": "Raid Volume", 00:07:26.210 "block_size": 512, 00:07:26.210 "num_blocks": 126976, 00:07:26.210 "uuid": "921efa90-8be7-4233-93dc-1f685a9600e7", 00:07:26.210 "assigned_rate_limits": { 00:07:26.210 "rw_ios_per_sec": 0, 00:07:26.210 "rw_mbytes_per_sec": 0, 00:07:26.210 "r_mbytes_per_sec": 0, 00:07:26.210 "w_mbytes_per_sec": 0 00:07:26.210 }, 00:07:26.210 "claimed": false, 00:07:26.210 "zoned": false, 00:07:26.210 "supported_io_types": { 00:07:26.210 "read": true, 00:07:26.210 "write": true, 00:07:26.210 "unmap": true, 00:07:26.210 "flush": true, 00:07:26.210 "reset": true, 00:07:26.210 "nvme_admin": false, 00:07:26.210 "nvme_io": false, 00:07:26.210 "nvme_io_md": false, 00:07:26.210 "write_zeroes": true, 00:07:26.210 "zcopy": false, 00:07:26.210 "get_zone_info": false, 00:07:26.210 "zone_management": false, 00:07:26.210 "zone_append": false, 00:07:26.210 "compare": false, 00:07:26.210 "compare_and_write": false, 00:07:26.210 "abort": false, 00:07:26.210 "seek_hole": false, 00:07:26.210 "seek_data": false, 00:07:26.210 "copy": false, 00:07:26.211 "nvme_iov_md": false 00:07:26.211 }, 00:07:26.211 "memory_domains": [ 00:07:26.211 { 00:07:26.211 "dma_device_id": "system", 00:07:26.211 "dma_device_type": 1 00:07:26.211 }, 00:07:26.211 { 00:07:26.211 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:26.211 "dma_device_type": 2 00:07:26.211 }, 00:07:26.211 { 00:07:26.211 "dma_device_id": "system", 00:07:26.211 "dma_device_type": 1 00:07:26.211 }, 00:07:26.211 { 00:07:26.211 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:26.211 "dma_device_type": 2 00:07:26.211 } 00:07:26.211 ], 00:07:26.211 "driver_specific": { 00:07:26.211 "raid": { 00:07:26.211 "uuid": "921efa90-8be7-4233-93dc-1f685a9600e7", 00:07:26.211 "strip_size_kb": 64, 00:07:26.211 "state": "online", 00:07:26.211 "raid_level": "raid0", 00:07:26.211 "superblock": true, 00:07:26.211 "num_base_bdevs": 2, 00:07:26.211 "num_base_bdevs_discovered": 2, 00:07:26.211 "num_base_bdevs_operational": 2, 00:07:26.211 "base_bdevs_list": [ 00:07:26.211 { 00:07:26.211 "name": "pt1", 00:07:26.211 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:26.211 "is_configured": true, 00:07:26.211 "data_offset": 2048, 00:07:26.211 "data_size": 63488 00:07:26.211 }, 00:07:26.211 { 00:07:26.211 "name": "pt2", 00:07:26.211 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:26.211 "is_configured": true, 00:07:26.211 "data_offset": 2048, 00:07:26.211 "data_size": 63488 00:07:26.211 } 00:07:26.211 ] 00:07:26.211 } 00:07:26.211 } 00:07:26.211 }' 00:07:26.211 15:11:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:26.211 15:11:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:26.211 pt2' 00:07:26.211 15:11:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:26.211 15:11:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:26.211 15:11:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:26.211 15:11:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:26.211 15:11:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:26.211 15:11:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.211 15:11:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:26.211 15:11:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:26.211 15:11:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:26.211 15:11:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:26.211 15:11:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:26.211 15:11:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:26.211 15:11:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:26.211 15:11:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.211 15:11:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:26.211 15:11:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:26.211 15:11:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:26.211 15:11:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:26.211 15:11:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:07:26.211 15:11:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:26.211 15:11:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:26.211 15:11:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.211 [2024-11-27 15:11:54.311378] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:26.471 15:11:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:26.471 15:11:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 921efa90-8be7-4233-93dc-1f685a9600e7 '!=' 921efa90-8be7-4233-93dc-1f685a9600e7 ']' 00:07:26.471 15:11:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:07:26.471 15:11:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:26.471 15:11:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:26.471 15:11:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 72757 00:07:26.471 15:11:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 72757 ']' 00:07:26.471 15:11:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 72757 00:07:26.471 15:11:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:07:26.471 15:11:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:26.471 15:11:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72757 00:07:26.471 15:11:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:26.471 15:11:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:26.471 killing process with pid 72757 00:07:26.471 15:11:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72757' 00:07:26.471 15:11:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 72757 00:07:26.471 [2024-11-27 15:11:54.388154] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:26.471 [2024-11-27 15:11:54.388248] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:26.471 [2024-11-27 15:11:54.388304] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:26.471 [2024-11-27 15:11:54.388319] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:07:26.471 15:11:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 72757 00:07:26.471 [2024-11-27 15:11:54.411851] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:26.730 15:11:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:07:26.730 00:07:26.730 real 0m3.292s 00:07:26.730 user 0m5.074s 00:07:26.730 sys 0m0.736s 00:07:26.730 15:11:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:26.731 15:11:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.731 ************************************ 00:07:26.731 END TEST raid_superblock_test 00:07:26.731 ************************************ 00:07:26.731 15:11:54 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 2 read 00:07:26.731 15:11:54 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:26.731 15:11:54 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:26.731 15:11:54 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:26.731 ************************************ 00:07:26.731 START TEST raid_read_error_test 00:07:26.731 ************************************ 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 2 read 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.lArSw7O4fe 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=72952 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 72952 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 72952 ']' 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:26.731 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:26.731 15:11:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.731 [2024-11-27 15:11:54.792774] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:07:26.731 [2024-11-27 15:11:54.792889] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72952 ] 00:07:26.991 [2024-11-27 15:11:54.962376] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:26.991 [2024-11-27 15:11:54.991258] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:26.991 [2024-11-27 15:11:55.035841] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:26.991 [2024-11-27 15:11:55.035884] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:27.560 15:11:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:27.560 15:11:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:07:27.560 15:11:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:27.560 15:11:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:27.560 15:11:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:27.560 15:11:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.560 BaseBdev1_malloc 00:07:27.560 15:11:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:27.560 15:11:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:27.560 15:11:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:27.560 15:11:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.560 true 00:07:27.560 15:11:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:27.560 15:11:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:27.560 15:11:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:27.560 15:11:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.819 [2024-11-27 15:11:55.668503] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:27.819 [2024-11-27 15:11:55.668562] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:27.819 [2024-11-27 15:11:55.668583] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:07:27.819 [2024-11-27 15:11:55.668592] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:27.819 [2024-11-27 15:11:55.670733] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:27.819 [2024-11-27 15:11:55.670766] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:27.819 BaseBdev1 00:07:27.819 15:11:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:27.819 15:11:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:27.819 15:11:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:27.819 15:11:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:27.819 15:11:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.819 BaseBdev2_malloc 00:07:27.819 15:11:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:27.819 15:11:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:27.819 15:11:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:27.819 15:11:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.819 true 00:07:27.819 15:11:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:27.819 15:11:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:27.819 15:11:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:27.819 15:11:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.820 [2024-11-27 15:11:55.709185] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:27.820 [2024-11-27 15:11:55.709239] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:27.820 [2024-11-27 15:11:55.709258] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:07:27.820 [2024-11-27 15:11:55.709266] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:27.820 [2024-11-27 15:11:55.711358] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:27.820 [2024-11-27 15:11:55.711399] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:27.820 BaseBdev2 00:07:27.820 15:11:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:27.820 15:11:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:27.820 15:11:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:27.820 15:11:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.820 [2024-11-27 15:11:55.721225] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:27.820 [2024-11-27 15:11:55.723157] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:27.820 [2024-11-27 15:11:55.723333] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:07:27.820 [2024-11-27 15:11:55.723347] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:27.820 [2024-11-27 15:11:55.723586] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:07:27.820 [2024-11-27 15:11:55.723720] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:07:27.820 [2024-11-27 15:11:55.723732] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:07:27.820 [2024-11-27 15:11:55.723895] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:27.820 15:11:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:27.820 15:11:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:27.820 15:11:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:27.820 15:11:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:27.820 15:11:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:27.820 15:11:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:27.820 15:11:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:27.820 15:11:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:27.820 15:11:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:27.820 15:11:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:27.820 15:11:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:27.820 15:11:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:27.820 15:11:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:27.820 15:11:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:27.820 15:11:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.820 15:11:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:27.820 15:11:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:27.820 "name": "raid_bdev1", 00:07:27.820 "uuid": "50295dac-a346-459a-be63-f5e45d124a67", 00:07:27.820 "strip_size_kb": 64, 00:07:27.820 "state": "online", 00:07:27.820 "raid_level": "raid0", 00:07:27.820 "superblock": true, 00:07:27.820 "num_base_bdevs": 2, 00:07:27.820 "num_base_bdevs_discovered": 2, 00:07:27.820 "num_base_bdevs_operational": 2, 00:07:27.820 "base_bdevs_list": [ 00:07:27.820 { 00:07:27.820 "name": "BaseBdev1", 00:07:27.820 "uuid": "6976dec3-b988-51d9-8406-8df646044d7c", 00:07:27.820 "is_configured": true, 00:07:27.820 "data_offset": 2048, 00:07:27.820 "data_size": 63488 00:07:27.820 }, 00:07:27.820 { 00:07:27.820 "name": "BaseBdev2", 00:07:27.820 "uuid": "2b71f920-13a5-5431-895e-dfea06fc4066", 00:07:27.820 "is_configured": true, 00:07:27.820 "data_offset": 2048, 00:07:27.820 "data_size": 63488 00:07:27.820 } 00:07:27.820 ] 00:07:27.820 }' 00:07:27.820 15:11:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:27.820 15:11:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:28.079 15:11:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:28.079 15:11:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:28.337 [2024-11-27 15:11:56.240724] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:07:29.273 15:11:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:07:29.273 15:11:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:29.273 15:11:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.273 15:11:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:29.273 15:11:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:29.273 15:11:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:07:29.273 15:11:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:07:29.273 15:11:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:29.273 15:11:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:29.273 15:11:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:29.273 15:11:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:29.273 15:11:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:29.273 15:11:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:29.273 15:11:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:29.273 15:11:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:29.273 15:11:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:29.273 15:11:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:29.273 15:11:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:29.274 15:11:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:29.274 15:11:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:29.274 15:11:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.274 15:11:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:29.274 15:11:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:29.274 "name": "raid_bdev1", 00:07:29.274 "uuid": "50295dac-a346-459a-be63-f5e45d124a67", 00:07:29.274 "strip_size_kb": 64, 00:07:29.274 "state": "online", 00:07:29.274 "raid_level": "raid0", 00:07:29.274 "superblock": true, 00:07:29.274 "num_base_bdevs": 2, 00:07:29.274 "num_base_bdevs_discovered": 2, 00:07:29.274 "num_base_bdevs_operational": 2, 00:07:29.274 "base_bdevs_list": [ 00:07:29.274 { 00:07:29.274 "name": "BaseBdev1", 00:07:29.274 "uuid": "6976dec3-b988-51d9-8406-8df646044d7c", 00:07:29.274 "is_configured": true, 00:07:29.274 "data_offset": 2048, 00:07:29.274 "data_size": 63488 00:07:29.274 }, 00:07:29.274 { 00:07:29.274 "name": "BaseBdev2", 00:07:29.274 "uuid": "2b71f920-13a5-5431-895e-dfea06fc4066", 00:07:29.274 "is_configured": true, 00:07:29.274 "data_offset": 2048, 00:07:29.274 "data_size": 63488 00:07:29.274 } 00:07:29.274 ] 00:07:29.274 }' 00:07:29.274 15:11:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:29.274 15:11:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.532 15:11:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:29.532 15:11:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:29.532 15:11:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.532 [2024-11-27 15:11:57.556410] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:29.532 [2024-11-27 15:11:57.556531] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:29.532 [2024-11-27 15:11:57.559098] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:29.532 [2024-11-27 15:11:57.559136] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:29.532 [2024-11-27 15:11:57.559169] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:29.532 [2024-11-27 15:11:57.559179] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:07:29.532 15:11:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:29.532 { 00:07:29.532 "results": [ 00:07:29.532 { 00:07:29.532 "job": "raid_bdev1", 00:07:29.532 "core_mask": "0x1", 00:07:29.532 "workload": "randrw", 00:07:29.532 "percentage": 50, 00:07:29.532 "status": "finished", 00:07:29.532 "queue_depth": 1, 00:07:29.532 "io_size": 131072, 00:07:29.532 "runtime": 1.316543, 00:07:29.532 "iops": 16792.46329212187, 00:07:29.532 "mibps": 2099.0579115152336, 00:07:29.532 "io_failed": 1, 00:07:29.532 "io_timeout": 0, 00:07:29.532 "avg_latency_us": 82.03419366651254, 00:07:29.532 "min_latency_us": 24.929257641921396, 00:07:29.532 "max_latency_us": 1359.3711790393013 00:07:29.532 } 00:07:29.532 ], 00:07:29.532 "core_count": 1 00:07:29.532 } 00:07:29.532 15:11:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 72952 00:07:29.532 15:11:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 72952 ']' 00:07:29.532 15:11:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 72952 00:07:29.532 15:11:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:07:29.532 15:11:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:29.532 15:11:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72952 00:07:29.532 15:11:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:29.532 15:11:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:29.533 15:11:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72952' 00:07:29.533 killing process with pid 72952 00:07:29.533 15:11:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 72952 00:07:29.533 [2024-11-27 15:11:57.606647] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:29.533 15:11:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 72952 00:07:29.533 [2024-11-27 15:11:57.622440] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:29.791 15:11:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.lArSw7O4fe 00:07:29.791 15:11:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:29.791 15:11:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:29.791 15:11:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.76 00:07:29.791 15:11:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:07:29.791 15:11:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:29.791 15:11:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:29.791 15:11:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.76 != \0\.\0\0 ]] 00:07:29.791 00:07:29.791 real 0m3.150s 00:07:29.791 user 0m4.002s 00:07:29.791 sys 0m0.490s 00:07:29.791 15:11:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:29.791 15:11:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.791 ************************************ 00:07:29.791 END TEST raid_read_error_test 00:07:29.791 ************************************ 00:07:30.049 15:11:57 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 2 write 00:07:30.049 15:11:57 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:30.049 15:11:57 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:30.049 15:11:57 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:30.049 ************************************ 00:07:30.049 START TEST raid_write_error_test 00:07:30.049 ************************************ 00:07:30.049 15:11:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 2 write 00:07:30.049 15:11:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:07:30.050 15:11:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:30.050 15:11:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:07:30.050 15:11:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:30.050 15:11:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:30.050 15:11:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:30.050 15:11:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:30.050 15:11:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:30.050 15:11:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:30.050 15:11:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:30.050 15:11:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:30.050 15:11:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:30.050 15:11:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:30.050 15:11:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:30.050 15:11:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:30.050 15:11:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:30.050 15:11:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:30.050 15:11:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:30.050 15:11:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:07:30.050 15:11:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:07:30.050 15:11:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:07:30.050 15:11:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:30.050 15:11:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.LX5AvJcCKP 00:07:30.050 15:11:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=73081 00:07:30.050 15:11:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:30.050 15:11:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 73081 00:07:30.050 15:11:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 73081 ']' 00:07:30.050 15:11:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:30.050 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:30.050 15:11:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:30.050 15:11:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:30.050 15:11:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:30.050 15:11:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.050 [2024-11-27 15:11:58.010459] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:07:30.050 [2024-11-27 15:11:58.010571] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73081 ] 00:07:30.308 [2024-11-27 15:11:58.180626] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:30.308 [2024-11-27 15:11:58.208717] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:30.308 [2024-11-27 15:11:58.253071] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:30.309 [2024-11-27 15:11:58.253104] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:30.876 15:11:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:30.876 15:11:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:07:30.876 15:11:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.877 BaseBdev1_malloc 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.877 true 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.877 [2024-11-27 15:11:58.870009] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:30.877 [2024-11-27 15:11:58.870073] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:30.877 [2024-11-27 15:11:58.870092] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:07:30.877 [2024-11-27 15:11:58.870101] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:30.877 [2024-11-27 15:11:58.872299] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:30.877 [2024-11-27 15:11:58.872419] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:30.877 BaseBdev1 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.877 BaseBdev2_malloc 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.877 true 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.877 [2024-11-27 15:11:58.910872] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:30.877 [2024-11-27 15:11:58.910958] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:30.877 [2024-11-27 15:11:58.910976] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:07:30.877 [2024-11-27 15:11:58.910985] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:30.877 [2024-11-27 15:11:58.913031] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:30.877 [2024-11-27 15:11:58.913085] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:30.877 BaseBdev2 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.877 [2024-11-27 15:11:58.922938] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:30.877 [2024-11-27 15:11:58.924740] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:30.877 [2024-11-27 15:11:58.924916] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:07:30.877 [2024-11-27 15:11:58.924931] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:30.877 [2024-11-27 15:11:58.925179] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:07:30.877 [2024-11-27 15:11:58.925319] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:07:30.877 [2024-11-27 15:11:58.925331] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:07:30.877 [2024-11-27 15:11:58.925455] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:30.877 15:11:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:30.877 "name": "raid_bdev1", 00:07:30.877 "uuid": "5be597be-236f-4785-87e4-1a6235697942", 00:07:30.877 "strip_size_kb": 64, 00:07:30.877 "state": "online", 00:07:30.877 "raid_level": "raid0", 00:07:30.877 "superblock": true, 00:07:30.877 "num_base_bdevs": 2, 00:07:30.877 "num_base_bdevs_discovered": 2, 00:07:30.877 "num_base_bdevs_operational": 2, 00:07:30.877 "base_bdevs_list": [ 00:07:30.878 { 00:07:30.878 "name": "BaseBdev1", 00:07:30.878 "uuid": "b9b402c3-cf68-527f-aafe-9dae1586470c", 00:07:30.878 "is_configured": true, 00:07:30.878 "data_offset": 2048, 00:07:30.878 "data_size": 63488 00:07:30.878 }, 00:07:30.878 { 00:07:30.878 "name": "BaseBdev2", 00:07:30.878 "uuid": "5f68041b-3234-556e-bab3-e73b8d44e420", 00:07:30.878 "is_configured": true, 00:07:30.878 "data_offset": 2048, 00:07:30.878 "data_size": 63488 00:07:30.878 } 00:07:30.878 ] 00:07:30.878 }' 00:07:30.878 15:11:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:30.878 15:11:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:31.445 15:11:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:31.445 15:11:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:31.445 [2024-11-27 15:11:59.466350] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:07:32.385 15:12:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:07:32.385 15:12:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:32.385 15:12:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.385 15:12:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:32.385 15:12:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:32.385 15:12:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:07:32.385 15:12:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:07:32.385 15:12:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:32.385 15:12:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:32.385 15:12:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:32.385 15:12:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:32.385 15:12:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:32.385 15:12:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:32.385 15:12:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:32.385 15:12:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:32.385 15:12:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:32.385 15:12:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:32.385 15:12:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:32.385 15:12:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:32.385 15:12:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:32.385 15:12:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.385 15:12:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:32.385 15:12:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:32.385 "name": "raid_bdev1", 00:07:32.385 "uuid": "5be597be-236f-4785-87e4-1a6235697942", 00:07:32.385 "strip_size_kb": 64, 00:07:32.385 "state": "online", 00:07:32.385 "raid_level": "raid0", 00:07:32.385 "superblock": true, 00:07:32.385 "num_base_bdevs": 2, 00:07:32.385 "num_base_bdevs_discovered": 2, 00:07:32.385 "num_base_bdevs_operational": 2, 00:07:32.385 "base_bdevs_list": [ 00:07:32.385 { 00:07:32.385 "name": "BaseBdev1", 00:07:32.385 "uuid": "b9b402c3-cf68-527f-aafe-9dae1586470c", 00:07:32.385 "is_configured": true, 00:07:32.385 "data_offset": 2048, 00:07:32.385 "data_size": 63488 00:07:32.385 }, 00:07:32.385 { 00:07:32.385 "name": "BaseBdev2", 00:07:32.385 "uuid": "5f68041b-3234-556e-bab3-e73b8d44e420", 00:07:32.385 "is_configured": true, 00:07:32.385 "data_offset": 2048, 00:07:32.385 "data_size": 63488 00:07:32.385 } 00:07:32.385 ] 00:07:32.385 }' 00:07:32.385 15:12:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:32.385 15:12:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.954 15:12:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:32.954 15:12:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:32.954 15:12:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.954 [2024-11-27 15:12:00.858193] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:32.954 [2024-11-27 15:12:00.858233] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:32.954 [2024-11-27 15:12:00.860782] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:32.954 [2024-11-27 15:12:00.860829] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:32.954 [2024-11-27 15:12:00.860864] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:32.954 [2024-11-27 15:12:00.860872] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:07:32.954 15:12:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:32.954 { 00:07:32.954 "results": [ 00:07:32.954 { 00:07:32.954 "job": "raid_bdev1", 00:07:32.954 "core_mask": "0x1", 00:07:32.954 "workload": "randrw", 00:07:32.954 "percentage": 50, 00:07:32.954 "status": "finished", 00:07:32.954 "queue_depth": 1, 00:07:32.954 "io_size": 131072, 00:07:32.954 "runtime": 1.392819, 00:07:32.954 "iops": 17211.85595544001, 00:07:32.954 "mibps": 2151.481994430001, 00:07:32.954 "io_failed": 1, 00:07:32.954 "io_timeout": 0, 00:07:32.954 "avg_latency_us": 80.05981093783186, 00:07:32.954 "min_latency_us": 24.817467248908297, 00:07:32.954 "max_latency_us": 1387.989519650655 00:07:32.954 } 00:07:32.954 ], 00:07:32.954 "core_count": 1 00:07:32.954 } 00:07:32.954 15:12:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 73081 00:07:32.954 15:12:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 73081 ']' 00:07:32.954 15:12:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 73081 00:07:32.954 15:12:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:07:32.954 15:12:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:32.954 15:12:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73081 00:07:32.954 15:12:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:32.954 15:12:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:32.954 15:12:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73081' 00:07:32.954 killing process with pid 73081 00:07:32.954 15:12:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 73081 00:07:32.954 [2024-11-27 15:12:00.908587] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:32.954 15:12:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 73081 00:07:32.954 [2024-11-27 15:12:00.923892] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:33.214 15:12:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:33.214 15:12:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.LX5AvJcCKP 00:07:33.214 15:12:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:33.214 15:12:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.72 00:07:33.214 15:12:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:07:33.214 15:12:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:33.214 15:12:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:33.214 15:12:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.72 != \0\.\0\0 ]] 00:07:33.214 00:07:33.214 real 0m3.226s 00:07:33.214 user 0m4.131s 00:07:33.214 sys 0m0.494s 00:07:33.214 15:12:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:33.214 15:12:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:33.214 ************************************ 00:07:33.214 END TEST raid_write_error_test 00:07:33.214 ************************************ 00:07:33.214 15:12:01 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:07:33.214 15:12:01 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 2 false 00:07:33.214 15:12:01 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:33.214 15:12:01 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:33.214 15:12:01 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:33.214 ************************************ 00:07:33.214 START TEST raid_state_function_test 00:07:33.214 ************************************ 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 2 false 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=73208 00:07:33.214 Process raid pid: 73208 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 73208' 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 73208 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 73208 ']' 00:07:33.214 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:33.214 15:12:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:33.214 [2024-11-27 15:12:01.293890] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:07:33.214 [2024-11-27 15:12:01.294016] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:33.473 [2024-11-27 15:12:01.464853] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:33.473 [2024-11-27 15:12:01.490084] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:33.473 [2024-11-27 15:12:01.533753] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:33.473 [2024-11-27 15:12:01.533797] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:34.040 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:34.040 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:07:34.040 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:34.040 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:34.040 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.040 [2024-11-27 15:12:02.137447] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:34.040 [2024-11-27 15:12:02.137511] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:34.040 [2024-11-27 15:12:02.137521] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:34.040 [2024-11-27 15:12:02.137548] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:34.040 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:34.040 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:34.040 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:34.040 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:34.040 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:34.041 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:34.041 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:34.041 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:34.041 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:34.041 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:34.041 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:34.299 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:34.299 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:34.299 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:34.299 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.299 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:34.299 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:34.299 "name": "Existed_Raid", 00:07:34.300 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:34.300 "strip_size_kb": 64, 00:07:34.300 "state": "configuring", 00:07:34.300 "raid_level": "concat", 00:07:34.300 "superblock": false, 00:07:34.300 "num_base_bdevs": 2, 00:07:34.300 "num_base_bdevs_discovered": 0, 00:07:34.300 "num_base_bdevs_operational": 2, 00:07:34.300 "base_bdevs_list": [ 00:07:34.300 { 00:07:34.300 "name": "BaseBdev1", 00:07:34.300 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:34.300 "is_configured": false, 00:07:34.300 "data_offset": 0, 00:07:34.300 "data_size": 0 00:07:34.300 }, 00:07:34.300 { 00:07:34.300 "name": "BaseBdev2", 00:07:34.300 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:34.300 "is_configured": false, 00:07:34.300 "data_offset": 0, 00:07:34.300 "data_size": 0 00:07:34.300 } 00:07:34.300 ] 00:07:34.300 }' 00:07:34.300 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:34.300 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.559 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:34.559 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:34.559 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.559 [2024-11-27 15:12:02.508738] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:34.559 [2024-11-27 15:12:02.508871] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:07:34.559 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:34.559 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:34.559 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:34.559 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.559 [2024-11-27 15:12:02.516721] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:34.559 [2024-11-27 15:12:02.516768] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:34.559 [2024-11-27 15:12:02.516776] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:34.559 [2024-11-27 15:12:02.516803] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:34.559 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:34.559 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:34.559 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:34.559 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.559 [2024-11-27 15:12:02.533750] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:34.559 BaseBdev1 00:07:34.559 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:34.559 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:34.559 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:07:34.559 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:34.559 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:34.559 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:34.559 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:34.559 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:34.559 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:34.560 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.560 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:34.560 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:34.560 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:34.560 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.560 [ 00:07:34.560 { 00:07:34.560 "name": "BaseBdev1", 00:07:34.560 "aliases": [ 00:07:34.560 "8df6c7f2-1724-43ff-8c8a-9affa1f7b905" 00:07:34.560 ], 00:07:34.560 "product_name": "Malloc disk", 00:07:34.560 "block_size": 512, 00:07:34.560 "num_blocks": 65536, 00:07:34.560 "uuid": "8df6c7f2-1724-43ff-8c8a-9affa1f7b905", 00:07:34.560 "assigned_rate_limits": { 00:07:34.560 "rw_ios_per_sec": 0, 00:07:34.560 "rw_mbytes_per_sec": 0, 00:07:34.560 "r_mbytes_per_sec": 0, 00:07:34.560 "w_mbytes_per_sec": 0 00:07:34.560 }, 00:07:34.560 "claimed": true, 00:07:34.560 "claim_type": "exclusive_write", 00:07:34.560 "zoned": false, 00:07:34.560 "supported_io_types": { 00:07:34.560 "read": true, 00:07:34.560 "write": true, 00:07:34.560 "unmap": true, 00:07:34.560 "flush": true, 00:07:34.560 "reset": true, 00:07:34.560 "nvme_admin": false, 00:07:34.560 "nvme_io": false, 00:07:34.560 "nvme_io_md": false, 00:07:34.560 "write_zeroes": true, 00:07:34.560 "zcopy": true, 00:07:34.560 "get_zone_info": false, 00:07:34.560 "zone_management": false, 00:07:34.560 "zone_append": false, 00:07:34.560 "compare": false, 00:07:34.560 "compare_and_write": false, 00:07:34.560 "abort": true, 00:07:34.560 "seek_hole": false, 00:07:34.560 "seek_data": false, 00:07:34.560 "copy": true, 00:07:34.560 "nvme_iov_md": false 00:07:34.560 }, 00:07:34.560 "memory_domains": [ 00:07:34.560 { 00:07:34.560 "dma_device_id": "system", 00:07:34.560 "dma_device_type": 1 00:07:34.560 }, 00:07:34.560 { 00:07:34.560 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:34.560 "dma_device_type": 2 00:07:34.560 } 00:07:34.560 ], 00:07:34.560 "driver_specific": {} 00:07:34.560 } 00:07:34.560 ] 00:07:34.560 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:34.560 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:34.560 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:34.560 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:34.560 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:34.560 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:34.560 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:34.560 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:34.560 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:34.560 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:34.560 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:34.560 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:34.560 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:34.560 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:34.560 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:34.560 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.560 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:34.560 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:34.560 "name": "Existed_Raid", 00:07:34.560 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:34.560 "strip_size_kb": 64, 00:07:34.560 "state": "configuring", 00:07:34.560 "raid_level": "concat", 00:07:34.560 "superblock": false, 00:07:34.560 "num_base_bdevs": 2, 00:07:34.560 "num_base_bdevs_discovered": 1, 00:07:34.560 "num_base_bdevs_operational": 2, 00:07:34.560 "base_bdevs_list": [ 00:07:34.560 { 00:07:34.560 "name": "BaseBdev1", 00:07:34.560 "uuid": "8df6c7f2-1724-43ff-8c8a-9affa1f7b905", 00:07:34.560 "is_configured": true, 00:07:34.560 "data_offset": 0, 00:07:34.560 "data_size": 65536 00:07:34.560 }, 00:07:34.560 { 00:07:34.560 "name": "BaseBdev2", 00:07:34.560 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:34.560 "is_configured": false, 00:07:34.560 "data_offset": 0, 00:07:34.560 "data_size": 0 00:07:34.560 } 00:07:34.560 ] 00:07:34.560 }' 00:07:34.560 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:34.560 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.129 15:12:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:35.129 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:35.129 15:12:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.129 [2024-11-27 15:12:03.000983] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:35.129 [2024-11-27 15:12:03.001034] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:07:35.129 15:12:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:35.129 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:35.129 15:12:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:35.129 15:12:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.129 [2024-11-27 15:12:03.009009] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:35.129 [2024-11-27 15:12:03.010794] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:35.129 [2024-11-27 15:12:03.010833] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:35.129 15:12:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:35.129 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:35.129 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:35.129 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:35.129 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:35.129 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:35.129 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:35.129 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:35.129 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:35.129 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:35.129 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:35.129 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:35.129 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:35.129 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:35.129 15:12:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:35.129 15:12:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.129 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:35.129 15:12:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:35.129 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:35.129 "name": "Existed_Raid", 00:07:35.129 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:35.129 "strip_size_kb": 64, 00:07:35.129 "state": "configuring", 00:07:35.129 "raid_level": "concat", 00:07:35.129 "superblock": false, 00:07:35.129 "num_base_bdevs": 2, 00:07:35.129 "num_base_bdevs_discovered": 1, 00:07:35.129 "num_base_bdevs_operational": 2, 00:07:35.129 "base_bdevs_list": [ 00:07:35.129 { 00:07:35.129 "name": "BaseBdev1", 00:07:35.129 "uuid": "8df6c7f2-1724-43ff-8c8a-9affa1f7b905", 00:07:35.129 "is_configured": true, 00:07:35.129 "data_offset": 0, 00:07:35.129 "data_size": 65536 00:07:35.129 }, 00:07:35.129 { 00:07:35.129 "name": "BaseBdev2", 00:07:35.129 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:35.129 "is_configured": false, 00:07:35.129 "data_offset": 0, 00:07:35.129 "data_size": 0 00:07:35.129 } 00:07:35.129 ] 00:07:35.129 }' 00:07:35.129 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:35.129 15:12:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.389 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:35.389 15:12:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:35.389 15:12:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.389 [2024-11-27 15:12:03.479416] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:35.389 [2024-11-27 15:12:03.479542] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:07:35.389 [2024-11-27 15:12:03.479569] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:07:35.389 [2024-11-27 15:12:03.479913] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:07:35.389 [2024-11-27 15:12:03.480115] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:07:35.389 [2024-11-27 15:12:03.480164] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:07:35.389 [2024-11-27 15:12:03.480426] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:35.389 BaseBdev2 00:07:35.389 15:12:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:35.389 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:35.389 15:12:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:07:35.389 15:12:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:35.389 15:12:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:35.389 15:12:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:35.389 15:12:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:35.389 15:12:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:35.389 15:12:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:35.389 15:12:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.649 15:12:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:35.649 15:12:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:35.649 15:12:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:35.649 15:12:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.649 [ 00:07:35.649 { 00:07:35.649 "name": "BaseBdev2", 00:07:35.649 "aliases": [ 00:07:35.649 "d608393a-d423-4923-a40e-82e15548ed88" 00:07:35.649 ], 00:07:35.649 "product_name": "Malloc disk", 00:07:35.649 "block_size": 512, 00:07:35.649 "num_blocks": 65536, 00:07:35.649 "uuid": "d608393a-d423-4923-a40e-82e15548ed88", 00:07:35.649 "assigned_rate_limits": { 00:07:35.649 "rw_ios_per_sec": 0, 00:07:35.649 "rw_mbytes_per_sec": 0, 00:07:35.649 "r_mbytes_per_sec": 0, 00:07:35.649 "w_mbytes_per_sec": 0 00:07:35.649 }, 00:07:35.649 "claimed": true, 00:07:35.649 "claim_type": "exclusive_write", 00:07:35.649 "zoned": false, 00:07:35.649 "supported_io_types": { 00:07:35.649 "read": true, 00:07:35.649 "write": true, 00:07:35.649 "unmap": true, 00:07:35.649 "flush": true, 00:07:35.649 "reset": true, 00:07:35.649 "nvme_admin": false, 00:07:35.649 "nvme_io": false, 00:07:35.649 "nvme_io_md": false, 00:07:35.649 "write_zeroes": true, 00:07:35.649 "zcopy": true, 00:07:35.649 "get_zone_info": false, 00:07:35.649 "zone_management": false, 00:07:35.650 "zone_append": false, 00:07:35.650 "compare": false, 00:07:35.650 "compare_and_write": false, 00:07:35.650 "abort": true, 00:07:35.650 "seek_hole": false, 00:07:35.650 "seek_data": false, 00:07:35.650 "copy": true, 00:07:35.650 "nvme_iov_md": false 00:07:35.650 }, 00:07:35.650 "memory_domains": [ 00:07:35.650 { 00:07:35.650 "dma_device_id": "system", 00:07:35.650 "dma_device_type": 1 00:07:35.650 }, 00:07:35.650 { 00:07:35.650 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:35.650 "dma_device_type": 2 00:07:35.650 } 00:07:35.650 ], 00:07:35.650 "driver_specific": {} 00:07:35.650 } 00:07:35.650 ] 00:07:35.650 15:12:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:35.650 15:12:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:35.650 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:35.650 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:35.650 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 2 00:07:35.650 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:35.650 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:35.650 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:35.650 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:35.650 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:35.650 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:35.650 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:35.650 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:35.650 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:35.650 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:35.650 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:35.650 15:12:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:35.650 15:12:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.650 15:12:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:35.650 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:35.650 "name": "Existed_Raid", 00:07:35.650 "uuid": "9ac4164a-30ac-4385-b0a7-4860316aed5a", 00:07:35.650 "strip_size_kb": 64, 00:07:35.650 "state": "online", 00:07:35.650 "raid_level": "concat", 00:07:35.650 "superblock": false, 00:07:35.650 "num_base_bdevs": 2, 00:07:35.650 "num_base_bdevs_discovered": 2, 00:07:35.650 "num_base_bdevs_operational": 2, 00:07:35.650 "base_bdevs_list": [ 00:07:35.650 { 00:07:35.650 "name": "BaseBdev1", 00:07:35.650 "uuid": "8df6c7f2-1724-43ff-8c8a-9affa1f7b905", 00:07:35.650 "is_configured": true, 00:07:35.650 "data_offset": 0, 00:07:35.650 "data_size": 65536 00:07:35.650 }, 00:07:35.650 { 00:07:35.650 "name": "BaseBdev2", 00:07:35.650 "uuid": "d608393a-d423-4923-a40e-82e15548ed88", 00:07:35.650 "is_configured": true, 00:07:35.650 "data_offset": 0, 00:07:35.650 "data_size": 65536 00:07:35.650 } 00:07:35.650 ] 00:07:35.650 }' 00:07:35.650 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:35.650 15:12:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.909 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:35.909 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:35.909 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:35.909 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:35.909 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:35.909 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:35.909 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:35.909 15:12:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:35.909 15:12:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:35.909 15:12:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.909 [2024-11-27 15:12:03.978994] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:35.909 15:12:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:35.909 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:35.909 "name": "Existed_Raid", 00:07:35.909 "aliases": [ 00:07:35.909 "9ac4164a-30ac-4385-b0a7-4860316aed5a" 00:07:35.909 ], 00:07:35.909 "product_name": "Raid Volume", 00:07:35.909 "block_size": 512, 00:07:35.909 "num_blocks": 131072, 00:07:35.909 "uuid": "9ac4164a-30ac-4385-b0a7-4860316aed5a", 00:07:35.909 "assigned_rate_limits": { 00:07:35.909 "rw_ios_per_sec": 0, 00:07:35.909 "rw_mbytes_per_sec": 0, 00:07:35.909 "r_mbytes_per_sec": 0, 00:07:35.909 "w_mbytes_per_sec": 0 00:07:35.909 }, 00:07:35.909 "claimed": false, 00:07:35.909 "zoned": false, 00:07:35.909 "supported_io_types": { 00:07:35.909 "read": true, 00:07:35.909 "write": true, 00:07:35.909 "unmap": true, 00:07:35.909 "flush": true, 00:07:35.909 "reset": true, 00:07:35.910 "nvme_admin": false, 00:07:35.910 "nvme_io": false, 00:07:35.910 "nvme_io_md": false, 00:07:35.910 "write_zeroes": true, 00:07:35.910 "zcopy": false, 00:07:35.910 "get_zone_info": false, 00:07:35.910 "zone_management": false, 00:07:35.910 "zone_append": false, 00:07:35.910 "compare": false, 00:07:35.910 "compare_and_write": false, 00:07:35.910 "abort": false, 00:07:35.910 "seek_hole": false, 00:07:35.910 "seek_data": false, 00:07:35.910 "copy": false, 00:07:35.910 "nvme_iov_md": false 00:07:35.910 }, 00:07:35.910 "memory_domains": [ 00:07:35.910 { 00:07:35.910 "dma_device_id": "system", 00:07:35.910 "dma_device_type": 1 00:07:35.910 }, 00:07:35.910 { 00:07:35.910 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:35.910 "dma_device_type": 2 00:07:35.910 }, 00:07:35.910 { 00:07:35.910 "dma_device_id": "system", 00:07:35.910 "dma_device_type": 1 00:07:35.910 }, 00:07:35.910 { 00:07:35.910 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:35.910 "dma_device_type": 2 00:07:35.910 } 00:07:35.910 ], 00:07:35.910 "driver_specific": { 00:07:35.910 "raid": { 00:07:35.910 "uuid": "9ac4164a-30ac-4385-b0a7-4860316aed5a", 00:07:35.910 "strip_size_kb": 64, 00:07:35.910 "state": "online", 00:07:35.910 "raid_level": "concat", 00:07:35.910 "superblock": false, 00:07:35.910 "num_base_bdevs": 2, 00:07:35.910 "num_base_bdevs_discovered": 2, 00:07:35.910 "num_base_bdevs_operational": 2, 00:07:35.910 "base_bdevs_list": [ 00:07:35.910 { 00:07:35.910 "name": "BaseBdev1", 00:07:35.910 "uuid": "8df6c7f2-1724-43ff-8c8a-9affa1f7b905", 00:07:35.910 "is_configured": true, 00:07:35.910 "data_offset": 0, 00:07:35.910 "data_size": 65536 00:07:35.910 }, 00:07:35.910 { 00:07:35.910 "name": "BaseBdev2", 00:07:35.910 "uuid": "d608393a-d423-4923-a40e-82e15548ed88", 00:07:35.910 "is_configured": true, 00:07:35.910 "data_offset": 0, 00:07:35.910 "data_size": 65536 00:07:35.910 } 00:07:35.910 ] 00:07:35.910 } 00:07:35.910 } 00:07:35.910 }' 00:07:35.910 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:36.170 BaseBdev2' 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.170 [2024-11-27 15:12:04.190362] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:36.170 [2024-11-27 15:12:04.190448] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:36.170 [2024-11-27 15:12:04.190530] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 1 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:36.170 "name": "Existed_Raid", 00:07:36.170 "uuid": "9ac4164a-30ac-4385-b0a7-4860316aed5a", 00:07:36.170 "strip_size_kb": 64, 00:07:36.170 "state": "offline", 00:07:36.170 "raid_level": "concat", 00:07:36.170 "superblock": false, 00:07:36.170 "num_base_bdevs": 2, 00:07:36.170 "num_base_bdevs_discovered": 1, 00:07:36.170 "num_base_bdevs_operational": 1, 00:07:36.170 "base_bdevs_list": [ 00:07:36.170 { 00:07:36.170 "name": null, 00:07:36.170 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:36.170 "is_configured": false, 00:07:36.170 "data_offset": 0, 00:07:36.170 "data_size": 65536 00:07:36.170 }, 00:07:36.170 { 00:07:36.170 "name": "BaseBdev2", 00:07:36.170 "uuid": "d608393a-d423-4923-a40e-82e15548ed88", 00:07:36.170 "is_configured": true, 00:07:36.170 "data_offset": 0, 00:07:36.170 "data_size": 65536 00:07:36.170 } 00:07:36.170 ] 00:07:36.170 }' 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:36.170 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.740 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:36.740 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:36.740 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:36.740 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:36.740 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.740 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:36.740 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:36.740 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:36.740 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:36.740 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:36.740 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:36.740 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.740 [2024-11-27 15:12:04.625437] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:36.740 [2024-11-27 15:12:04.625507] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:07:36.740 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:36.740 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:36.740 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:36.740 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:36.740 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:36.740 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:36.740 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.740 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:36.740 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:36.740 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:36.740 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:36.740 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 73208 00:07:36.740 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 73208 ']' 00:07:36.741 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 73208 00:07:36.741 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:07:36.741 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:36.741 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73208 00:07:36.741 killing process with pid 73208 00:07:36.741 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:36.741 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:36.741 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73208' 00:07:36.741 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 73208 00:07:36.741 [2024-11-27 15:12:04.730233] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:36.741 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 73208 00:07:36.741 [2024-11-27 15:12:04.731181] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:37.000 15:12:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:07:37.000 00:07:37.000 real 0m3.749s 00:07:37.000 user 0m5.917s 00:07:37.000 sys 0m0.741s 00:07:37.000 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:37.000 15:12:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.000 ************************************ 00:07:37.000 END TEST raid_state_function_test 00:07:37.000 ************************************ 00:07:37.000 15:12:05 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 2 true 00:07:37.000 15:12:05 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:37.000 15:12:05 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:37.000 15:12:05 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:37.000 ************************************ 00:07:37.000 START TEST raid_state_function_test_sb 00:07:37.000 ************************************ 00:07:37.000 15:12:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 2 true 00:07:37.000 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:07:37.000 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:37.000 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:07:37.000 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:37.000 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:37.000 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:37.000 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:37.000 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:37.000 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:37.000 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:37.000 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:37.000 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:37.000 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:37.000 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:37.000 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:37.000 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:37.000 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:37.000 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:37.000 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:07:37.000 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:07:37.000 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:07:37.000 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:07:37.000 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:07:37.000 Process raid pid: 73450 00:07:37.000 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=73450 00:07:37.000 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:37.000 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 73450' 00:07:37.000 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 73450 00:07:37.001 15:12:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 73450 ']' 00:07:37.001 15:12:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:37.001 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:37.001 15:12:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:37.001 15:12:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:37.001 15:12:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:37.001 15:12:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:37.260 [2024-11-27 15:12:05.107305] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:07:37.260 [2024-11-27 15:12:05.107504] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:37.260 [2024-11-27 15:12:05.279809] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:37.260 [2024-11-27 15:12:05.306550] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:37.260 [2024-11-27 15:12:05.349839] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:37.260 [2024-11-27 15:12:05.349979] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:38.197 15:12:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:38.197 15:12:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:07:38.198 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:38.198 15:12:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:38.198 15:12:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:38.198 [2024-11-27 15:12:05.965716] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:38.198 [2024-11-27 15:12:05.965785] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:38.198 [2024-11-27 15:12:05.965812] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:38.198 [2024-11-27 15:12:05.965821] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:38.198 15:12:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:38.198 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:38.198 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:38.198 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:38.198 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:38.198 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:38.198 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:38.198 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:38.198 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:38.198 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:38.198 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:38.198 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:38.198 15:12:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:38.198 15:12:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:38.198 15:12:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:38.198 15:12:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:38.198 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:38.198 "name": "Existed_Raid", 00:07:38.198 "uuid": "9cd6ba5a-5681-4f25-8e86-c8ceb410b413", 00:07:38.198 "strip_size_kb": 64, 00:07:38.198 "state": "configuring", 00:07:38.198 "raid_level": "concat", 00:07:38.198 "superblock": true, 00:07:38.198 "num_base_bdevs": 2, 00:07:38.198 "num_base_bdevs_discovered": 0, 00:07:38.198 "num_base_bdevs_operational": 2, 00:07:38.198 "base_bdevs_list": [ 00:07:38.198 { 00:07:38.198 "name": "BaseBdev1", 00:07:38.198 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:38.198 "is_configured": false, 00:07:38.198 "data_offset": 0, 00:07:38.198 "data_size": 0 00:07:38.198 }, 00:07:38.198 { 00:07:38.198 "name": "BaseBdev2", 00:07:38.198 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:38.198 "is_configured": false, 00:07:38.198 "data_offset": 0, 00:07:38.198 "data_size": 0 00:07:38.198 } 00:07:38.198 ] 00:07:38.198 }' 00:07:38.198 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:38.198 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:38.457 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:38.457 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:38.457 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:38.457 [2024-11-27 15:12:06.376920] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:38.457 [2024-11-27 15:12:06.377038] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:07:38.457 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:38.457 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:38.457 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:38.457 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:38.457 [2024-11-27 15:12:06.388922] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:38.457 [2024-11-27 15:12:06.389009] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:38.457 [2024-11-27 15:12:06.389035] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:38.457 [2024-11-27 15:12:06.389059] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:38.457 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:38.457 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:38.457 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:38.457 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:38.457 [2024-11-27 15:12:06.409871] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:38.457 BaseBdev1 00:07:38.457 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:38.457 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:38.457 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:07:38.457 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:38.457 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:38.457 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:38.457 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:38.457 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:38.457 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:38.457 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:38.457 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:38.457 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:38.457 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:38.457 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:38.457 [ 00:07:38.457 { 00:07:38.457 "name": "BaseBdev1", 00:07:38.457 "aliases": [ 00:07:38.457 "a9276e41-5a50-4545-8548-129c86e62107" 00:07:38.457 ], 00:07:38.457 "product_name": "Malloc disk", 00:07:38.457 "block_size": 512, 00:07:38.457 "num_blocks": 65536, 00:07:38.457 "uuid": "a9276e41-5a50-4545-8548-129c86e62107", 00:07:38.457 "assigned_rate_limits": { 00:07:38.457 "rw_ios_per_sec": 0, 00:07:38.457 "rw_mbytes_per_sec": 0, 00:07:38.457 "r_mbytes_per_sec": 0, 00:07:38.457 "w_mbytes_per_sec": 0 00:07:38.457 }, 00:07:38.457 "claimed": true, 00:07:38.457 "claim_type": "exclusive_write", 00:07:38.457 "zoned": false, 00:07:38.457 "supported_io_types": { 00:07:38.457 "read": true, 00:07:38.457 "write": true, 00:07:38.457 "unmap": true, 00:07:38.457 "flush": true, 00:07:38.457 "reset": true, 00:07:38.457 "nvme_admin": false, 00:07:38.457 "nvme_io": false, 00:07:38.457 "nvme_io_md": false, 00:07:38.457 "write_zeroes": true, 00:07:38.457 "zcopy": true, 00:07:38.457 "get_zone_info": false, 00:07:38.457 "zone_management": false, 00:07:38.457 "zone_append": false, 00:07:38.457 "compare": false, 00:07:38.457 "compare_and_write": false, 00:07:38.457 "abort": true, 00:07:38.457 "seek_hole": false, 00:07:38.457 "seek_data": false, 00:07:38.457 "copy": true, 00:07:38.457 "nvme_iov_md": false 00:07:38.457 }, 00:07:38.457 "memory_domains": [ 00:07:38.457 { 00:07:38.457 "dma_device_id": "system", 00:07:38.457 "dma_device_type": 1 00:07:38.457 }, 00:07:38.457 { 00:07:38.457 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:38.457 "dma_device_type": 2 00:07:38.457 } 00:07:38.457 ], 00:07:38.457 "driver_specific": {} 00:07:38.457 } 00:07:38.457 ] 00:07:38.457 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:38.458 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:38.458 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:38.458 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:38.458 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:38.458 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:38.458 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:38.458 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:38.458 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:38.458 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:38.458 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:38.458 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:38.458 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:38.458 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:38.458 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:38.458 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:38.458 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:38.458 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:38.458 "name": "Existed_Raid", 00:07:38.458 "uuid": "5658dc24-4477-4860-9b39-94326b32df98", 00:07:38.458 "strip_size_kb": 64, 00:07:38.458 "state": "configuring", 00:07:38.458 "raid_level": "concat", 00:07:38.458 "superblock": true, 00:07:38.458 "num_base_bdevs": 2, 00:07:38.458 "num_base_bdevs_discovered": 1, 00:07:38.458 "num_base_bdevs_operational": 2, 00:07:38.458 "base_bdevs_list": [ 00:07:38.458 { 00:07:38.458 "name": "BaseBdev1", 00:07:38.458 "uuid": "a9276e41-5a50-4545-8548-129c86e62107", 00:07:38.458 "is_configured": true, 00:07:38.458 "data_offset": 2048, 00:07:38.458 "data_size": 63488 00:07:38.458 }, 00:07:38.458 { 00:07:38.458 "name": "BaseBdev2", 00:07:38.458 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:38.458 "is_configured": false, 00:07:38.458 "data_offset": 0, 00:07:38.458 "data_size": 0 00:07:38.458 } 00:07:38.458 ] 00:07:38.458 }' 00:07:38.458 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:38.458 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.027 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:39.027 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:39.027 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.027 [2024-11-27 15:12:06.893076] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:39.027 [2024-11-27 15:12:06.893204] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:07:39.027 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:39.027 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:39.027 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:39.027 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.027 [2024-11-27 15:12:06.905095] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:39.027 [2024-11-27 15:12:06.906937] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:39.027 [2024-11-27 15:12:06.907032] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:39.027 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:39.027 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:39.027 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:39.027 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:39.027 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:39.027 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:39.027 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:39.027 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:39.027 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:39.027 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:39.027 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:39.027 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:39.027 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:39.027 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:39.027 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:39.027 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:39.027 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.027 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:39.027 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:39.027 "name": "Existed_Raid", 00:07:39.027 "uuid": "df9751f9-ac63-474b-a1b5-665d78ae2721", 00:07:39.027 "strip_size_kb": 64, 00:07:39.027 "state": "configuring", 00:07:39.027 "raid_level": "concat", 00:07:39.027 "superblock": true, 00:07:39.027 "num_base_bdevs": 2, 00:07:39.027 "num_base_bdevs_discovered": 1, 00:07:39.027 "num_base_bdevs_operational": 2, 00:07:39.027 "base_bdevs_list": [ 00:07:39.027 { 00:07:39.027 "name": "BaseBdev1", 00:07:39.027 "uuid": "a9276e41-5a50-4545-8548-129c86e62107", 00:07:39.027 "is_configured": true, 00:07:39.027 "data_offset": 2048, 00:07:39.027 "data_size": 63488 00:07:39.027 }, 00:07:39.027 { 00:07:39.027 "name": "BaseBdev2", 00:07:39.027 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:39.027 "is_configured": false, 00:07:39.027 "data_offset": 0, 00:07:39.027 "data_size": 0 00:07:39.027 } 00:07:39.027 ] 00:07:39.027 }' 00:07:39.027 15:12:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:39.027 15:12:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.287 [2024-11-27 15:12:07.327593] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:39.287 [2024-11-27 15:12:07.327945] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:07:39.287 [2024-11-27 15:12:07.328002] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:39.287 BaseBdev2 00:07:39.287 [2024-11-27 15:12:07.328311] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:07:39.287 [2024-11-27 15:12:07.328470] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:07:39.287 [2024-11-27 15:12:07.328542] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:07:39.287 [2024-11-27 15:12:07.328703] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.287 [ 00:07:39.287 { 00:07:39.287 "name": "BaseBdev2", 00:07:39.287 "aliases": [ 00:07:39.287 "d8545f9f-ff90-4f58-b544-3e7451e3932d" 00:07:39.287 ], 00:07:39.287 "product_name": "Malloc disk", 00:07:39.287 "block_size": 512, 00:07:39.287 "num_blocks": 65536, 00:07:39.287 "uuid": "d8545f9f-ff90-4f58-b544-3e7451e3932d", 00:07:39.287 "assigned_rate_limits": { 00:07:39.287 "rw_ios_per_sec": 0, 00:07:39.287 "rw_mbytes_per_sec": 0, 00:07:39.287 "r_mbytes_per_sec": 0, 00:07:39.287 "w_mbytes_per_sec": 0 00:07:39.287 }, 00:07:39.287 "claimed": true, 00:07:39.287 "claim_type": "exclusive_write", 00:07:39.287 "zoned": false, 00:07:39.287 "supported_io_types": { 00:07:39.287 "read": true, 00:07:39.287 "write": true, 00:07:39.287 "unmap": true, 00:07:39.287 "flush": true, 00:07:39.287 "reset": true, 00:07:39.287 "nvme_admin": false, 00:07:39.287 "nvme_io": false, 00:07:39.287 "nvme_io_md": false, 00:07:39.287 "write_zeroes": true, 00:07:39.287 "zcopy": true, 00:07:39.287 "get_zone_info": false, 00:07:39.287 "zone_management": false, 00:07:39.287 "zone_append": false, 00:07:39.287 "compare": false, 00:07:39.287 "compare_and_write": false, 00:07:39.287 "abort": true, 00:07:39.287 "seek_hole": false, 00:07:39.287 "seek_data": false, 00:07:39.287 "copy": true, 00:07:39.287 "nvme_iov_md": false 00:07:39.287 }, 00:07:39.287 "memory_domains": [ 00:07:39.287 { 00:07:39.287 "dma_device_id": "system", 00:07:39.287 "dma_device_type": 1 00:07:39.287 }, 00:07:39.287 { 00:07:39.287 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:39.287 "dma_device_type": 2 00:07:39.287 } 00:07:39.287 ], 00:07:39.287 "driver_specific": {} 00:07:39.287 } 00:07:39.287 ] 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 2 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.287 15:12:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:39.547 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:39.547 "name": "Existed_Raid", 00:07:39.547 "uuid": "df9751f9-ac63-474b-a1b5-665d78ae2721", 00:07:39.547 "strip_size_kb": 64, 00:07:39.547 "state": "online", 00:07:39.547 "raid_level": "concat", 00:07:39.547 "superblock": true, 00:07:39.547 "num_base_bdevs": 2, 00:07:39.547 "num_base_bdevs_discovered": 2, 00:07:39.547 "num_base_bdevs_operational": 2, 00:07:39.547 "base_bdevs_list": [ 00:07:39.547 { 00:07:39.547 "name": "BaseBdev1", 00:07:39.547 "uuid": "a9276e41-5a50-4545-8548-129c86e62107", 00:07:39.547 "is_configured": true, 00:07:39.547 "data_offset": 2048, 00:07:39.547 "data_size": 63488 00:07:39.547 }, 00:07:39.547 { 00:07:39.547 "name": "BaseBdev2", 00:07:39.547 "uuid": "d8545f9f-ff90-4f58-b544-3e7451e3932d", 00:07:39.547 "is_configured": true, 00:07:39.547 "data_offset": 2048, 00:07:39.547 "data_size": 63488 00:07:39.547 } 00:07:39.547 ] 00:07:39.547 }' 00:07:39.547 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:39.547 15:12:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.805 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:39.805 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:39.805 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:39.805 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:39.805 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:07:39.805 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:39.805 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:39.805 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:39.805 15:12:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:39.805 15:12:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.806 [2024-11-27 15:12:07.815148] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:39.806 15:12:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:39.806 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:39.806 "name": "Existed_Raid", 00:07:39.806 "aliases": [ 00:07:39.806 "df9751f9-ac63-474b-a1b5-665d78ae2721" 00:07:39.806 ], 00:07:39.806 "product_name": "Raid Volume", 00:07:39.806 "block_size": 512, 00:07:39.806 "num_blocks": 126976, 00:07:39.806 "uuid": "df9751f9-ac63-474b-a1b5-665d78ae2721", 00:07:39.806 "assigned_rate_limits": { 00:07:39.806 "rw_ios_per_sec": 0, 00:07:39.806 "rw_mbytes_per_sec": 0, 00:07:39.806 "r_mbytes_per_sec": 0, 00:07:39.806 "w_mbytes_per_sec": 0 00:07:39.806 }, 00:07:39.806 "claimed": false, 00:07:39.806 "zoned": false, 00:07:39.806 "supported_io_types": { 00:07:39.806 "read": true, 00:07:39.806 "write": true, 00:07:39.806 "unmap": true, 00:07:39.806 "flush": true, 00:07:39.806 "reset": true, 00:07:39.806 "nvme_admin": false, 00:07:39.806 "nvme_io": false, 00:07:39.806 "nvme_io_md": false, 00:07:39.806 "write_zeroes": true, 00:07:39.806 "zcopy": false, 00:07:39.806 "get_zone_info": false, 00:07:39.806 "zone_management": false, 00:07:39.806 "zone_append": false, 00:07:39.806 "compare": false, 00:07:39.806 "compare_and_write": false, 00:07:39.806 "abort": false, 00:07:39.806 "seek_hole": false, 00:07:39.806 "seek_data": false, 00:07:39.806 "copy": false, 00:07:39.806 "nvme_iov_md": false 00:07:39.806 }, 00:07:39.806 "memory_domains": [ 00:07:39.806 { 00:07:39.806 "dma_device_id": "system", 00:07:39.806 "dma_device_type": 1 00:07:39.806 }, 00:07:39.806 { 00:07:39.806 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:39.806 "dma_device_type": 2 00:07:39.806 }, 00:07:39.806 { 00:07:39.806 "dma_device_id": "system", 00:07:39.806 "dma_device_type": 1 00:07:39.806 }, 00:07:39.806 { 00:07:39.806 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:39.806 "dma_device_type": 2 00:07:39.806 } 00:07:39.806 ], 00:07:39.806 "driver_specific": { 00:07:39.806 "raid": { 00:07:39.806 "uuid": "df9751f9-ac63-474b-a1b5-665d78ae2721", 00:07:39.806 "strip_size_kb": 64, 00:07:39.806 "state": "online", 00:07:39.806 "raid_level": "concat", 00:07:39.806 "superblock": true, 00:07:39.806 "num_base_bdevs": 2, 00:07:39.806 "num_base_bdevs_discovered": 2, 00:07:39.806 "num_base_bdevs_operational": 2, 00:07:39.806 "base_bdevs_list": [ 00:07:39.806 { 00:07:39.806 "name": "BaseBdev1", 00:07:39.806 "uuid": "a9276e41-5a50-4545-8548-129c86e62107", 00:07:39.806 "is_configured": true, 00:07:39.806 "data_offset": 2048, 00:07:39.806 "data_size": 63488 00:07:39.806 }, 00:07:39.806 { 00:07:39.806 "name": "BaseBdev2", 00:07:39.806 "uuid": "d8545f9f-ff90-4f58-b544-3e7451e3932d", 00:07:39.806 "is_configured": true, 00:07:39.806 "data_offset": 2048, 00:07:39.806 "data_size": 63488 00:07:39.806 } 00:07:39.806 ] 00:07:39.806 } 00:07:39.806 } 00:07:39.806 }' 00:07:39.806 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:39.806 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:39.806 BaseBdev2' 00:07:39.806 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:40.065 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:40.065 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:40.065 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:40.065 15:12:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.065 15:12:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.065 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:40.065 15:12:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.065 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:40.065 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:40.065 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:40.066 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:40.066 15:12:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.066 15:12:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:40.066 15:12:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.066 15:12:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.066 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:40.066 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:40.066 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:40.066 15:12:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.066 15:12:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.066 [2024-11-27 15:12:08.038474] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:40.066 [2024-11-27 15:12:08.038554] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:40.066 [2024-11-27 15:12:08.038639] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:40.066 15:12:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.066 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:40.066 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:07:40.066 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:40.066 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:07:40.066 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:40.066 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 1 00:07:40.066 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:40.066 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:40.066 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:40.066 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:40.066 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:40.066 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:40.066 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:40.066 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:40.066 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:40.066 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:40.066 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:40.066 15:12:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.066 15:12:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.066 15:12:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.066 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:40.066 "name": "Existed_Raid", 00:07:40.066 "uuid": "df9751f9-ac63-474b-a1b5-665d78ae2721", 00:07:40.066 "strip_size_kb": 64, 00:07:40.066 "state": "offline", 00:07:40.066 "raid_level": "concat", 00:07:40.066 "superblock": true, 00:07:40.066 "num_base_bdevs": 2, 00:07:40.066 "num_base_bdevs_discovered": 1, 00:07:40.066 "num_base_bdevs_operational": 1, 00:07:40.066 "base_bdevs_list": [ 00:07:40.066 { 00:07:40.066 "name": null, 00:07:40.066 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:40.066 "is_configured": false, 00:07:40.066 "data_offset": 0, 00:07:40.066 "data_size": 63488 00:07:40.066 }, 00:07:40.066 { 00:07:40.066 "name": "BaseBdev2", 00:07:40.066 "uuid": "d8545f9f-ff90-4f58-b544-3e7451e3932d", 00:07:40.066 "is_configured": true, 00:07:40.066 "data_offset": 2048, 00:07:40.066 "data_size": 63488 00:07:40.066 } 00:07:40.066 ] 00:07:40.066 }' 00:07:40.066 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:40.066 15:12:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.633 [2024-11-27 15:12:08.545076] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:40.633 [2024-11-27 15:12:08.545200] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 73450 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 73450 ']' 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 73450 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73450 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73450' 00:07:40.633 killing process with pid 73450 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 73450 00:07:40.633 [2024-11-27 15:12:08.638065] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:40.633 15:12:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 73450 00:07:40.633 [2024-11-27 15:12:08.639035] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:40.893 15:12:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:07:40.893 00:07:40.893 real 0m3.831s 00:07:40.893 user 0m6.090s 00:07:40.893 sys 0m0.746s 00:07:40.893 ************************************ 00:07:40.893 END TEST raid_state_function_test_sb 00:07:40.893 ************************************ 00:07:40.893 15:12:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:40.893 15:12:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.893 15:12:08 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 2 00:07:40.893 15:12:08 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:07:40.893 15:12:08 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:40.893 15:12:08 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:40.893 ************************************ 00:07:40.893 START TEST raid_superblock_test 00:07:40.893 ************************************ 00:07:40.893 15:12:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test concat 2 00:07:40.893 15:12:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:07:40.893 15:12:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:07:40.893 15:12:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:07:40.893 15:12:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:07:40.893 15:12:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:07:40.893 15:12:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:07:40.893 15:12:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:07:40.893 15:12:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:07:40.893 15:12:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:07:40.893 15:12:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:07:40.893 15:12:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:07:40.893 15:12:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:07:40.893 15:12:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:07:40.893 15:12:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:07:40.893 15:12:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:07:40.893 15:12:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:07:40.893 15:12:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:07:40.893 15:12:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=73691 00:07:40.893 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:40.893 15:12:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 73691 00:07:40.893 15:12:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 73691 ']' 00:07:40.893 15:12:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:40.893 15:12:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:40.893 15:12:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:40.893 15:12:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:40.893 15:12:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:41.154 [2024-11-27 15:12:09.005643] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:07:41.154 [2024-11-27 15:12:09.005777] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73691 ] 00:07:41.154 [2024-11-27 15:12:09.173043] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:41.154 [2024-11-27 15:12:09.199145] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:41.154 [2024-11-27 15:12:09.243680] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:41.154 [2024-11-27 15:12:09.243729] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:41.723 15:12:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:41.723 15:12:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:07:41.723 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:07:41.723 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:41.723 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:07:41.723 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:07:41.723 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:07:41.723 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:41.723 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:41.723 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:41.723 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:07:41.983 15:12:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:41.983 15:12:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:41.983 malloc1 00:07:41.983 15:12:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:41.983 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:41.983 15:12:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:41.984 [2024-11-27 15:12:09.853096] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:41.984 [2024-11-27 15:12:09.853223] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:41.984 [2024-11-27 15:12:09.853269] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:07:41.984 [2024-11-27 15:12:09.853304] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:41.984 [2024-11-27 15:12:09.855418] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:41.984 [2024-11-27 15:12:09.855494] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:41.984 pt1 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:41.984 malloc2 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:41.984 [2024-11-27 15:12:09.885952] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:41.984 [2024-11-27 15:12:09.886007] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:41.984 [2024-11-27 15:12:09.886023] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:07:41.984 [2024-11-27 15:12:09.886033] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:41.984 [2024-11-27 15:12:09.888107] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:41.984 [2024-11-27 15:12:09.888146] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:41.984 pt2 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:41.984 [2024-11-27 15:12:09.897969] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:41.984 [2024-11-27 15:12:09.899809] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:41.984 [2024-11-27 15:12:09.899980] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:07:41.984 [2024-11-27 15:12:09.899997] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:41.984 [2024-11-27 15:12:09.900238] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:07:41.984 [2024-11-27 15:12:09.900355] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:07:41.984 [2024-11-27 15:12:09.900372] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:07:41.984 [2024-11-27 15:12:09.900492] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:41.984 "name": "raid_bdev1", 00:07:41.984 "uuid": "6419c3a9-41a2-456d-a4c2-8f45a05948f7", 00:07:41.984 "strip_size_kb": 64, 00:07:41.984 "state": "online", 00:07:41.984 "raid_level": "concat", 00:07:41.984 "superblock": true, 00:07:41.984 "num_base_bdevs": 2, 00:07:41.984 "num_base_bdevs_discovered": 2, 00:07:41.984 "num_base_bdevs_operational": 2, 00:07:41.984 "base_bdevs_list": [ 00:07:41.984 { 00:07:41.984 "name": "pt1", 00:07:41.984 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:41.984 "is_configured": true, 00:07:41.984 "data_offset": 2048, 00:07:41.984 "data_size": 63488 00:07:41.984 }, 00:07:41.984 { 00:07:41.984 "name": "pt2", 00:07:41.984 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:41.984 "is_configured": true, 00:07:41.984 "data_offset": 2048, 00:07:41.984 "data_size": 63488 00:07:41.984 } 00:07:41.984 ] 00:07:41.984 }' 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:41.984 15:12:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.553 [2024-11-27 15:12:10.357437] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:42.553 "name": "raid_bdev1", 00:07:42.553 "aliases": [ 00:07:42.553 "6419c3a9-41a2-456d-a4c2-8f45a05948f7" 00:07:42.553 ], 00:07:42.553 "product_name": "Raid Volume", 00:07:42.553 "block_size": 512, 00:07:42.553 "num_blocks": 126976, 00:07:42.553 "uuid": "6419c3a9-41a2-456d-a4c2-8f45a05948f7", 00:07:42.553 "assigned_rate_limits": { 00:07:42.553 "rw_ios_per_sec": 0, 00:07:42.553 "rw_mbytes_per_sec": 0, 00:07:42.553 "r_mbytes_per_sec": 0, 00:07:42.553 "w_mbytes_per_sec": 0 00:07:42.553 }, 00:07:42.553 "claimed": false, 00:07:42.553 "zoned": false, 00:07:42.553 "supported_io_types": { 00:07:42.553 "read": true, 00:07:42.553 "write": true, 00:07:42.553 "unmap": true, 00:07:42.553 "flush": true, 00:07:42.553 "reset": true, 00:07:42.553 "nvme_admin": false, 00:07:42.553 "nvme_io": false, 00:07:42.553 "nvme_io_md": false, 00:07:42.553 "write_zeroes": true, 00:07:42.553 "zcopy": false, 00:07:42.553 "get_zone_info": false, 00:07:42.553 "zone_management": false, 00:07:42.553 "zone_append": false, 00:07:42.553 "compare": false, 00:07:42.553 "compare_and_write": false, 00:07:42.553 "abort": false, 00:07:42.553 "seek_hole": false, 00:07:42.553 "seek_data": false, 00:07:42.553 "copy": false, 00:07:42.553 "nvme_iov_md": false 00:07:42.553 }, 00:07:42.553 "memory_domains": [ 00:07:42.553 { 00:07:42.553 "dma_device_id": "system", 00:07:42.553 "dma_device_type": 1 00:07:42.553 }, 00:07:42.553 { 00:07:42.553 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:42.553 "dma_device_type": 2 00:07:42.553 }, 00:07:42.553 { 00:07:42.553 "dma_device_id": "system", 00:07:42.553 "dma_device_type": 1 00:07:42.553 }, 00:07:42.553 { 00:07:42.553 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:42.553 "dma_device_type": 2 00:07:42.553 } 00:07:42.553 ], 00:07:42.553 "driver_specific": { 00:07:42.553 "raid": { 00:07:42.553 "uuid": "6419c3a9-41a2-456d-a4c2-8f45a05948f7", 00:07:42.553 "strip_size_kb": 64, 00:07:42.553 "state": "online", 00:07:42.553 "raid_level": "concat", 00:07:42.553 "superblock": true, 00:07:42.553 "num_base_bdevs": 2, 00:07:42.553 "num_base_bdevs_discovered": 2, 00:07:42.553 "num_base_bdevs_operational": 2, 00:07:42.553 "base_bdevs_list": [ 00:07:42.553 { 00:07:42.553 "name": "pt1", 00:07:42.553 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:42.553 "is_configured": true, 00:07:42.553 "data_offset": 2048, 00:07:42.553 "data_size": 63488 00:07:42.553 }, 00:07:42.553 { 00:07:42.553 "name": "pt2", 00:07:42.553 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:42.553 "is_configured": true, 00:07:42.553 "data_offset": 2048, 00:07:42.553 "data_size": 63488 00:07:42.553 } 00:07:42.553 ] 00:07:42.553 } 00:07:42.553 } 00:07:42.553 }' 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:42.553 pt2' 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.553 [2024-11-27 15:12:10.560994] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=6419c3a9-41a2-456d-a4c2-8f45a05948f7 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 6419c3a9-41a2-456d-a4c2-8f45a05948f7 ']' 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.553 [2024-11-27 15:12:10.596697] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:42.553 [2024-11-27 15:12:10.596775] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:42.553 [2024-11-27 15:12:10.596856] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:42.553 [2024-11-27 15:12:10.596922] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:42.553 [2024-11-27 15:12:10.596936] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:42.553 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.814 [2024-11-27 15:12:10.736493] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:07:42.814 [2024-11-27 15:12:10.738406] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:07:42.814 [2024-11-27 15:12:10.738476] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:07:42.814 [2024-11-27 15:12:10.738521] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:07:42.814 [2024-11-27 15:12:10.738538] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:42.814 [2024-11-27 15:12:10.738547] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state configuring 00:07:42.814 request: 00:07:42.814 { 00:07:42.814 "name": "raid_bdev1", 00:07:42.814 "raid_level": "concat", 00:07:42.814 "base_bdevs": [ 00:07:42.814 "malloc1", 00:07:42.814 "malloc2" 00:07:42.814 ], 00:07:42.814 "strip_size_kb": 64, 00:07:42.814 "superblock": false, 00:07:42.814 "method": "bdev_raid_create", 00:07:42.814 "req_id": 1 00:07:42.814 } 00:07:42.814 Got JSON-RPC error response 00:07:42.814 response: 00:07:42.814 { 00:07:42.814 "code": -17, 00:07:42.814 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:07:42.814 } 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.814 [2024-11-27 15:12:10.800322] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:42.814 [2024-11-27 15:12:10.800434] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:42.814 [2024-11-27 15:12:10.800474] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:07:42.814 [2024-11-27 15:12:10.800503] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:42.814 [2024-11-27 15:12:10.802627] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:42.814 [2024-11-27 15:12:10.802696] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:42.814 [2024-11-27 15:12:10.802793] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:07:42.814 [2024-11-27 15:12:10.802842] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:42.814 pt1 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 2 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:42.814 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:42.815 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:42.815 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:42.815 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:42.815 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:42.815 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:42.815 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:42.815 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.815 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:42.815 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:42.815 "name": "raid_bdev1", 00:07:42.815 "uuid": "6419c3a9-41a2-456d-a4c2-8f45a05948f7", 00:07:42.815 "strip_size_kb": 64, 00:07:42.815 "state": "configuring", 00:07:42.815 "raid_level": "concat", 00:07:42.815 "superblock": true, 00:07:42.815 "num_base_bdevs": 2, 00:07:42.815 "num_base_bdevs_discovered": 1, 00:07:42.815 "num_base_bdevs_operational": 2, 00:07:42.815 "base_bdevs_list": [ 00:07:42.815 { 00:07:42.815 "name": "pt1", 00:07:42.815 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:42.815 "is_configured": true, 00:07:42.815 "data_offset": 2048, 00:07:42.815 "data_size": 63488 00:07:42.815 }, 00:07:42.815 { 00:07:42.815 "name": null, 00:07:42.815 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:42.815 "is_configured": false, 00:07:42.815 "data_offset": 2048, 00:07:42.815 "data_size": 63488 00:07:42.815 } 00:07:42.815 ] 00:07:42.815 }' 00:07:42.815 15:12:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:42.815 15:12:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.075 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:07:43.075 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:07:43.075 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:43.075 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:43.075 15:12:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:43.075 15:12:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.075 [2024-11-27 15:12:11.171957] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:43.075 [2024-11-27 15:12:11.172028] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:43.075 [2024-11-27 15:12:11.172053] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:07:43.075 [2024-11-27 15:12:11.172062] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:43.075 [2024-11-27 15:12:11.172490] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:43.075 [2024-11-27 15:12:11.172508] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:43.075 [2024-11-27 15:12:11.172585] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:07:43.075 [2024-11-27 15:12:11.172605] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:43.075 [2024-11-27 15:12:11.172702] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:07:43.075 [2024-11-27 15:12:11.172712] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:43.075 [2024-11-27 15:12:11.172957] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:07:43.075 [2024-11-27 15:12:11.173073] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:07:43.075 [2024-11-27 15:12:11.173086] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:07:43.075 [2024-11-27 15:12:11.173186] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:43.075 pt2 00:07:43.076 15:12:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:43.076 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:07:43.076 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:43.076 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:43.076 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:43.076 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:43.076 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:43.076 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:43.076 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:43.076 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:43.076 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:43.076 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:43.076 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:43.335 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:43.335 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:43.335 15:12:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:43.336 15:12:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.336 15:12:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:43.336 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:43.336 "name": "raid_bdev1", 00:07:43.336 "uuid": "6419c3a9-41a2-456d-a4c2-8f45a05948f7", 00:07:43.336 "strip_size_kb": 64, 00:07:43.336 "state": "online", 00:07:43.336 "raid_level": "concat", 00:07:43.336 "superblock": true, 00:07:43.336 "num_base_bdevs": 2, 00:07:43.336 "num_base_bdevs_discovered": 2, 00:07:43.336 "num_base_bdevs_operational": 2, 00:07:43.336 "base_bdevs_list": [ 00:07:43.336 { 00:07:43.336 "name": "pt1", 00:07:43.336 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:43.336 "is_configured": true, 00:07:43.336 "data_offset": 2048, 00:07:43.336 "data_size": 63488 00:07:43.336 }, 00:07:43.336 { 00:07:43.336 "name": "pt2", 00:07:43.336 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:43.336 "is_configured": true, 00:07:43.336 "data_offset": 2048, 00:07:43.336 "data_size": 63488 00:07:43.336 } 00:07:43.336 ] 00:07:43.336 }' 00:07:43.336 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:43.336 15:12:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.596 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:07:43.596 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:43.596 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:43.596 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:43.596 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:43.596 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:43.596 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:43.596 15:12:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:43.596 15:12:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.596 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:43.596 [2024-11-27 15:12:11.595596] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:43.596 15:12:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:43.596 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:43.596 "name": "raid_bdev1", 00:07:43.596 "aliases": [ 00:07:43.596 "6419c3a9-41a2-456d-a4c2-8f45a05948f7" 00:07:43.596 ], 00:07:43.596 "product_name": "Raid Volume", 00:07:43.596 "block_size": 512, 00:07:43.596 "num_blocks": 126976, 00:07:43.596 "uuid": "6419c3a9-41a2-456d-a4c2-8f45a05948f7", 00:07:43.596 "assigned_rate_limits": { 00:07:43.596 "rw_ios_per_sec": 0, 00:07:43.596 "rw_mbytes_per_sec": 0, 00:07:43.596 "r_mbytes_per_sec": 0, 00:07:43.596 "w_mbytes_per_sec": 0 00:07:43.596 }, 00:07:43.596 "claimed": false, 00:07:43.596 "zoned": false, 00:07:43.596 "supported_io_types": { 00:07:43.596 "read": true, 00:07:43.596 "write": true, 00:07:43.596 "unmap": true, 00:07:43.596 "flush": true, 00:07:43.596 "reset": true, 00:07:43.596 "nvme_admin": false, 00:07:43.596 "nvme_io": false, 00:07:43.596 "nvme_io_md": false, 00:07:43.596 "write_zeroes": true, 00:07:43.596 "zcopy": false, 00:07:43.596 "get_zone_info": false, 00:07:43.596 "zone_management": false, 00:07:43.596 "zone_append": false, 00:07:43.596 "compare": false, 00:07:43.596 "compare_and_write": false, 00:07:43.596 "abort": false, 00:07:43.596 "seek_hole": false, 00:07:43.596 "seek_data": false, 00:07:43.596 "copy": false, 00:07:43.596 "nvme_iov_md": false 00:07:43.596 }, 00:07:43.596 "memory_domains": [ 00:07:43.596 { 00:07:43.596 "dma_device_id": "system", 00:07:43.596 "dma_device_type": 1 00:07:43.596 }, 00:07:43.596 { 00:07:43.596 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:43.596 "dma_device_type": 2 00:07:43.596 }, 00:07:43.596 { 00:07:43.596 "dma_device_id": "system", 00:07:43.596 "dma_device_type": 1 00:07:43.596 }, 00:07:43.596 { 00:07:43.596 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:43.596 "dma_device_type": 2 00:07:43.596 } 00:07:43.596 ], 00:07:43.596 "driver_specific": { 00:07:43.596 "raid": { 00:07:43.596 "uuid": "6419c3a9-41a2-456d-a4c2-8f45a05948f7", 00:07:43.596 "strip_size_kb": 64, 00:07:43.596 "state": "online", 00:07:43.596 "raid_level": "concat", 00:07:43.596 "superblock": true, 00:07:43.596 "num_base_bdevs": 2, 00:07:43.596 "num_base_bdevs_discovered": 2, 00:07:43.596 "num_base_bdevs_operational": 2, 00:07:43.596 "base_bdevs_list": [ 00:07:43.596 { 00:07:43.596 "name": "pt1", 00:07:43.596 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:43.596 "is_configured": true, 00:07:43.596 "data_offset": 2048, 00:07:43.596 "data_size": 63488 00:07:43.596 }, 00:07:43.596 { 00:07:43.596 "name": "pt2", 00:07:43.596 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:43.596 "is_configured": true, 00:07:43.596 "data_offset": 2048, 00:07:43.596 "data_size": 63488 00:07:43.596 } 00:07:43.596 ] 00:07:43.596 } 00:07:43.596 } 00:07:43.596 }' 00:07:43.596 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:43.596 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:43.596 pt2' 00:07:43.596 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:43.596 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:43.596 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:43.596 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:43.596 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:43.596 15:12:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:43.596 15:12:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.856 15:12:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:43.856 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:43.856 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:43.856 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:43.856 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:43.856 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:43.856 15:12:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:43.856 15:12:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.856 15:12:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:43.856 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:43.856 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:43.856 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:43.856 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:07:43.856 15:12:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:43.856 15:12:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.856 [2024-11-27 15:12:11.775245] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:43.856 15:12:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:43.856 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 6419c3a9-41a2-456d-a4c2-8f45a05948f7 '!=' 6419c3a9-41a2-456d-a4c2-8f45a05948f7 ']' 00:07:43.856 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:07:43.857 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:43.857 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:43.857 15:12:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 73691 00:07:43.857 15:12:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 73691 ']' 00:07:43.857 15:12:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 73691 00:07:43.857 15:12:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:07:43.857 15:12:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:43.857 15:12:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73691 00:07:43.857 killing process with pid 73691 00:07:43.857 15:12:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:43.857 15:12:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:43.857 15:12:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73691' 00:07:43.857 15:12:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 73691 00:07:43.857 [2024-11-27 15:12:11.859878] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:43.857 [2024-11-27 15:12:11.859972] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:43.857 [2024-11-27 15:12:11.860027] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:43.857 [2024-11-27 15:12:11.860037] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:07:43.857 15:12:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 73691 00:07:43.857 [2024-11-27 15:12:11.883629] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:44.116 15:12:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:07:44.116 00:07:44.116 real 0m3.177s 00:07:44.116 user 0m4.885s 00:07:44.116 sys 0m0.685s 00:07:44.116 15:12:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:44.116 15:12:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.116 ************************************ 00:07:44.116 END TEST raid_superblock_test 00:07:44.116 ************************************ 00:07:44.116 15:12:12 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 2 read 00:07:44.116 15:12:12 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:44.116 15:12:12 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:44.116 15:12:12 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:44.116 ************************************ 00:07:44.116 START TEST raid_read_error_test 00:07:44.116 ************************************ 00:07:44.116 15:12:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 2 read 00:07:44.116 15:12:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:07:44.117 15:12:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:44.117 15:12:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:07:44.117 15:12:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:44.117 15:12:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:44.117 15:12:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:44.117 15:12:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:44.117 15:12:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:44.117 15:12:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:44.117 15:12:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:44.117 15:12:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:44.117 15:12:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:44.117 15:12:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:44.117 15:12:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:44.117 15:12:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:44.117 15:12:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:44.117 15:12:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:44.117 15:12:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:44.117 15:12:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:07:44.117 15:12:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:07:44.117 15:12:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:07:44.117 15:12:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:44.117 15:12:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.agFZCsGZI6 00:07:44.117 15:12:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=73886 00:07:44.117 15:12:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 73886 00:07:44.117 15:12:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:44.117 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:44.117 15:12:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 73886 ']' 00:07:44.117 15:12:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:44.117 15:12:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:44.117 15:12:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:44.117 15:12:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:44.117 15:12:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.376 [2024-11-27 15:12:12.255847] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:07:44.376 [2024-11-27 15:12:12.256073] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73886 ] 00:07:44.376 [2024-11-27 15:12:12.411279] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:44.376 [2024-11-27 15:12:12.436570] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:44.376 [2024-11-27 15:12:12.480494] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:44.376 [2024-11-27 15:12:12.480627] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.316 BaseBdev1_malloc 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.316 true 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.316 [2024-11-27 15:12:13.113390] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:45.316 [2024-11-27 15:12:13.113452] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:45.316 [2024-11-27 15:12:13.113499] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:07:45.316 [2024-11-27 15:12:13.113508] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:45.316 [2024-11-27 15:12:13.115567] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:45.316 [2024-11-27 15:12:13.115685] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:45.316 BaseBdev1 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.316 BaseBdev2_malloc 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.316 true 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.316 [2024-11-27 15:12:13.154170] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:45.316 [2024-11-27 15:12:13.154223] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:45.316 [2024-11-27 15:12:13.154241] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:07:45.316 [2024-11-27 15:12:13.154249] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:45.316 [2024-11-27 15:12:13.156363] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:45.316 [2024-11-27 15:12:13.156408] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:45.316 BaseBdev2 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.316 [2024-11-27 15:12:13.166209] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:45.316 [2024-11-27 15:12:13.168025] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:45.316 [2024-11-27 15:12:13.168212] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:07:45.316 [2024-11-27 15:12:13.168225] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:45.316 [2024-11-27 15:12:13.168471] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:07:45.316 [2024-11-27 15:12:13.168608] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:07:45.316 [2024-11-27 15:12:13.168620] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:07:45.316 [2024-11-27 15:12:13.168735] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:45.316 "name": "raid_bdev1", 00:07:45.316 "uuid": "cb55cfe2-5eb1-42ce-a490-16a193a717c4", 00:07:45.316 "strip_size_kb": 64, 00:07:45.316 "state": "online", 00:07:45.316 "raid_level": "concat", 00:07:45.316 "superblock": true, 00:07:45.316 "num_base_bdevs": 2, 00:07:45.316 "num_base_bdevs_discovered": 2, 00:07:45.316 "num_base_bdevs_operational": 2, 00:07:45.316 "base_bdevs_list": [ 00:07:45.316 { 00:07:45.316 "name": "BaseBdev1", 00:07:45.316 "uuid": "11786dcf-ac67-5e75-a290-91b9121b4b91", 00:07:45.316 "is_configured": true, 00:07:45.316 "data_offset": 2048, 00:07:45.316 "data_size": 63488 00:07:45.316 }, 00:07:45.316 { 00:07:45.316 "name": "BaseBdev2", 00:07:45.316 "uuid": "4a04e0d0-d4ff-59d6-8c16-07b85c0b2002", 00:07:45.316 "is_configured": true, 00:07:45.316 "data_offset": 2048, 00:07:45.316 "data_size": 63488 00:07:45.316 } 00:07:45.316 ] 00:07:45.316 }' 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:45.316 15:12:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.575 15:12:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:45.575 15:12:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:45.834 [2024-11-27 15:12:13.713623] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:07:46.773 15:12:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:07:46.773 15:12:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:46.773 15:12:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.773 15:12:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:46.773 15:12:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:46.773 15:12:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:07:46.773 15:12:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:07:46.773 15:12:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:46.773 15:12:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:46.773 15:12:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:46.774 15:12:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:46.774 15:12:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:46.774 15:12:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:46.774 15:12:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:46.774 15:12:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:46.774 15:12:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:46.774 15:12:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:46.774 15:12:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:46.774 15:12:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:46.774 15:12:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:46.774 15:12:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.774 15:12:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:46.774 15:12:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:46.774 "name": "raid_bdev1", 00:07:46.774 "uuid": "cb55cfe2-5eb1-42ce-a490-16a193a717c4", 00:07:46.774 "strip_size_kb": 64, 00:07:46.774 "state": "online", 00:07:46.774 "raid_level": "concat", 00:07:46.774 "superblock": true, 00:07:46.774 "num_base_bdevs": 2, 00:07:46.774 "num_base_bdevs_discovered": 2, 00:07:46.774 "num_base_bdevs_operational": 2, 00:07:46.774 "base_bdevs_list": [ 00:07:46.774 { 00:07:46.774 "name": "BaseBdev1", 00:07:46.774 "uuid": "11786dcf-ac67-5e75-a290-91b9121b4b91", 00:07:46.774 "is_configured": true, 00:07:46.774 "data_offset": 2048, 00:07:46.774 "data_size": 63488 00:07:46.774 }, 00:07:46.774 { 00:07:46.774 "name": "BaseBdev2", 00:07:46.774 "uuid": "4a04e0d0-d4ff-59d6-8c16-07b85c0b2002", 00:07:46.774 "is_configured": true, 00:07:46.774 "data_offset": 2048, 00:07:46.774 "data_size": 63488 00:07:46.774 } 00:07:46.774 ] 00:07:46.774 }' 00:07:46.774 15:12:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:46.774 15:12:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.034 15:12:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:47.034 15:12:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:47.034 15:12:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.034 [2024-11-27 15:12:15.061064] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:47.034 [2024-11-27 15:12:15.061183] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:47.034 [2024-11-27 15:12:15.064207] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:47.034 [2024-11-27 15:12:15.064293] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:47.034 [2024-11-27 15:12:15.064354] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:47.034 [2024-11-27 15:12:15.064416] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:07:47.034 { 00:07:47.034 "results": [ 00:07:47.034 { 00:07:47.034 "job": "raid_bdev1", 00:07:47.034 "core_mask": "0x1", 00:07:47.034 "workload": "randrw", 00:07:47.034 "percentage": 50, 00:07:47.034 "status": "finished", 00:07:47.034 "queue_depth": 1, 00:07:47.034 "io_size": 131072, 00:07:47.034 "runtime": 1.348521, 00:07:47.034 "iops": 16891.097728548535, 00:07:47.034 "mibps": 2111.387216068567, 00:07:47.034 "io_failed": 1, 00:07:47.034 "io_timeout": 0, 00:07:47.034 "avg_latency_us": 81.58697858346892, 00:07:47.034 "min_latency_us": 24.817467248908297, 00:07:47.034 "max_latency_us": 1366.5257641921398 00:07:47.034 } 00:07:47.034 ], 00:07:47.034 "core_count": 1 00:07:47.034 } 00:07:47.034 15:12:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:47.034 15:12:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 73886 00:07:47.034 15:12:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 73886 ']' 00:07:47.034 15:12:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 73886 00:07:47.034 15:12:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:07:47.034 15:12:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:47.034 15:12:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73886 00:07:47.034 15:12:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:47.034 15:12:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:47.034 15:12:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73886' 00:07:47.034 killing process with pid 73886 00:07:47.034 15:12:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 73886 00:07:47.034 [2024-11-27 15:12:15.112961] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:47.034 15:12:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 73886 00:07:47.034 [2024-11-27 15:12:15.128955] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:47.299 15:12:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:47.299 15:12:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.agFZCsGZI6 00:07:47.299 15:12:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:47.299 15:12:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.74 00:07:47.299 15:12:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:07:47.299 15:12:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:47.299 15:12:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:47.299 ************************************ 00:07:47.299 END TEST raid_read_error_test 00:07:47.299 ************************************ 00:07:47.299 15:12:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.74 != \0\.\0\0 ]] 00:07:47.299 00:07:47.299 real 0m3.178s 00:07:47.299 user 0m4.082s 00:07:47.299 sys 0m0.493s 00:07:47.299 15:12:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:47.299 15:12:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.299 15:12:15 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 2 write 00:07:47.299 15:12:15 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:47.299 15:12:15 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:47.299 15:12:15 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:47.568 ************************************ 00:07:47.568 START TEST raid_write_error_test 00:07:47.568 ************************************ 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 2 write 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.2uRwSVVZgt 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=74015 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 74015 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 74015 ']' 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:47.568 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:47.568 15:12:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.568 [2024-11-27 15:12:15.515393] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:07:47.568 [2024-11-27 15:12:15.515646] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid74015 ] 00:07:47.827 [2024-11-27 15:12:15.692541] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:47.827 [2024-11-27 15:12:15.718606] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:47.827 [2024-11-27 15:12:15.762123] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:47.827 [2024-11-27 15:12:15.762245] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:48.398 15:12:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:48.398 15:12:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:07:48.398 15:12:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:48.398 15:12:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:48.398 15:12:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:48.398 15:12:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.398 BaseBdev1_malloc 00:07:48.398 15:12:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:48.398 15:12:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:48.398 15:12:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:48.398 15:12:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.398 true 00:07:48.398 15:12:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:48.398 15:12:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:48.398 15:12:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:48.398 15:12:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.398 [2024-11-27 15:12:16.366896] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:48.398 [2024-11-27 15:12:16.367056] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:48.398 [2024-11-27 15:12:16.367088] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:07:48.398 [2024-11-27 15:12:16.367097] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:48.398 [2024-11-27 15:12:16.369274] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:48.398 [2024-11-27 15:12:16.369313] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:48.398 BaseBdev1 00:07:48.398 15:12:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:48.398 15:12:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:48.398 15:12:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:48.398 15:12:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:48.398 15:12:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.398 BaseBdev2_malloc 00:07:48.398 15:12:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:48.398 15:12:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:48.398 15:12:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:48.398 15:12:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.398 true 00:07:48.398 15:12:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:48.398 15:12:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:48.398 15:12:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:48.398 15:12:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.398 [2024-11-27 15:12:16.407738] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:48.398 [2024-11-27 15:12:16.407820] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:48.398 [2024-11-27 15:12:16.407842] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:07:48.398 [2024-11-27 15:12:16.407853] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:48.398 [2024-11-27 15:12:16.409926] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:48.398 [2024-11-27 15:12:16.409961] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:48.398 BaseBdev2 00:07:48.398 15:12:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:48.399 15:12:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:48.399 15:12:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:48.399 15:12:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.399 [2024-11-27 15:12:16.419821] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:48.399 [2024-11-27 15:12:16.421661] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:48.399 [2024-11-27 15:12:16.421892] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:07:48.399 [2024-11-27 15:12:16.421919] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:48.399 [2024-11-27 15:12:16.422162] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:07:48.399 [2024-11-27 15:12:16.422295] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:07:48.399 [2024-11-27 15:12:16.422313] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:07:48.399 [2024-11-27 15:12:16.422431] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:48.399 15:12:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:48.399 15:12:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:48.399 15:12:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:48.399 15:12:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:48.399 15:12:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:48.399 15:12:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:48.399 15:12:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:48.399 15:12:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:48.399 15:12:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:48.399 15:12:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:48.399 15:12:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:48.399 15:12:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:48.399 15:12:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:48.399 15:12:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.399 15:12:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:48.399 15:12:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:48.399 15:12:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:48.399 "name": "raid_bdev1", 00:07:48.399 "uuid": "9ad79f34-9188-49e2-947f-76af4871da59", 00:07:48.399 "strip_size_kb": 64, 00:07:48.399 "state": "online", 00:07:48.399 "raid_level": "concat", 00:07:48.399 "superblock": true, 00:07:48.399 "num_base_bdevs": 2, 00:07:48.399 "num_base_bdevs_discovered": 2, 00:07:48.399 "num_base_bdevs_operational": 2, 00:07:48.399 "base_bdevs_list": [ 00:07:48.399 { 00:07:48.399 "name": "BaseBdev1", 00:07:48.399 "uuid": "4e47196b-7e9c-5368-9d8a-5bb37589955c", 00:07:48.399 "is_configured": true, 00:07:48.399 "data_offset": 2048, 00:07:48.399 "data_size": 63488 00:07:48.399 }, 00:07:48.399 { 00:07:48.399 "name": "BaseBdev2", 00:07:48.399 "uuid": "709653c5-45bd-5a7d-8849-c9250800f24a", 00:07:48.399 "is_configured": true, 00:07:48.399 "data_offset": 2048, 00:07:48.399 "data_size": 63488 00:07:48.399 } 00:07:48.399 ] 00:07:48.399 }' 00:07:48.399 15:12:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:48.399 15:12:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.968 15:12:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:48.968 15:12:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:48.968 [2024-11-27 15:12:16.951253] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:07:49.907 15:12:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:07:49.907 15:12:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:49.907 15:12:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.907 15:12:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:49.907 15:12:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:49.907 15:12:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:07:49.907 15:12:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:07:49.907 15:12:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:49.907 15:12:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:49.907 15:12:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:49.907 15:12:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:49.907 15:12:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:49.907 15:12:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:49.907 15:12:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:49.907 15:12:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:49.907 15:12:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:49.907 15:12:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:49.907 15:12:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:49.907 15:12:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:49.907 15:12:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.907 15:12:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:49.907 15:12:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:49.907 15:12:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:49.907 "name": "raid_bdev1", 00:07:49.907 "uuid": "9ad79f34-9188-49e2-947f-76af4871da59", 00:07:49.907 "strip_size_kb": 64, 00:07:49.907 "state": "online", 00:07:49.907 "raid_level": "concat", 00:07:49.907 "superblock": true, 00:07:49.907 "num_base_bdevs": 2, 00:07:49.907 "num_base_bdevs_discovered": 2, 00:07:49.907 "num_base_bdevs_operational": 2, 00:07:49.907 "base_bdevs_list": [ 00:07:49.907 { 00:07:49.907 "name": "BaseBdev1", 00:07:49.907 "uuid": "4e47196b-7e9c-5368-9d8a-5bb37589955c", 00:07:49.907 "is_configured": true, 00:07:49.907 "data_offset": 2048, 00:07:49.907 "data_size": 63488 00:07:49.907 }, 00:07:49.907 { 00:07:49.907 "name": "BaseBdev2", 00:07:49.907 "uuid": "709653c5-45bd-5a7d-8849-c9250800f24a", 00:07:49.907 "is_configured": true, 00:07:49.907 "data_offset": 2048, 00:07:49.907 "data_size": 63488 00:07:49.907 } 00:07:49.907 ] 00:07:49.907 }' 00:07:49.907 15:12:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:49.907 15:12:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.477 15:12:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:50.477 15:12:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:50.477 15:12:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.477 [2024-11-27 15:12:18.339157] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:50.477 [2024-11-27 15:12:18.339194] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:50.477 [2024-11-27 15:12:18.341723] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:50.477 [2024-11-27 15:12:18.341767] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:50.477 [2024-11-27 15:12:18.341803] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:50.477 [2024-11-27 15:12:18.341812] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:07:50.477 { 00:07:50.477 "results": [ 00:07:50.477 { 00:07:50.477 "job": "raid_bdev1", 00:07:50.477 "core_mask": "0x1", 00:07:50.477 "workload": "randrw", 00:07:50.477 "percentage": 50, 00:07:50.477 "status": "finished", 00:07:50.477 "queue_depth": 1, 00:07:50.477 "io_size": 131072, 00:07:50.477 "runtime": 1.388806, 00:07:50.477 "iops": 17064.29839732835, 00:07:50.477 "mibps": 2133.037299666044, 00:07:50.477 "io_failed": 1, 00:07:50.477 "io_timeout": 0, 00:07:50.477 "avg_latency_us": 80.79986026200874, 00:07:50.477 "min_latency_us": 24.705676855895195, 00:07:50.477 "max_latency_us": 1402.2986899563318 00:07:50.477 } 00:07:50.477 ], 00:07:50.477 "core_count": 1 00:07:50.477 } 00:07:50.477 15:12:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:50.477 15:12:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 74015 00:07:50.477 15:12:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 74015 ']' 00:07:50.477 15:12:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 74015 00:07:50.477 15:12:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:07:50.477 15:12:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:50.477 15:12:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 74015 00:07:50.477 killing process with pid 74015 00:07:50.477 15:12:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:50.477 15:12:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:50.477 15:12:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 74015' 00:07:50.477 15:12:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 74015 00:07:50.477 [2024-11-27 15:12:18.388735] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:50.477 15:12:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 74015 00:07:50.477 [2024-11-27 15:12:18.404370] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:50.737 15:12:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:50.737 15:12:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.2uRwSVVZgt 00:07:50.737 15:12:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:50.737 15:12:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.72 00:07:50.737 15:12:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:07:50.737 ************************************ 00:07:50.737 END TEST raid_write_error_test 00:07:50.737 ************************************ 00:07:50.737 15:12:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:50.737 15:12:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:50.737 15:12:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.72 != \0\.\0\0 ]] 00:07:50.737 00:07:50.737 real 0m3.201s 00:07:50.737 user 0m4.083s 00:07:50.737 sys 0m0.508s 00:07:50.737 15:12:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:50.737 15:12:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.737 15:12:18 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:07:50.737 15:12:18 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 2 false 00:07:50.737 15:12:18 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:50.737 15:12:18 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:50.737 15:12:18 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:50.737 ************************************ 00:07:50.737 START TEST raid_state_function_test 00:07:50.737 ************************************ 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 false 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=74142 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 74142' 00:07:50.737 Process raid pid: 74142 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 74142 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 74142 ']' 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:50.737 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:50.737 15:12:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.737 [2024-11-27 15:12:18.771321] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:07:50.737 [2024-11-27 15:12:18.771550] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:50.997 [2024-11-27 15:12:18.945786] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:50.997 [2024-11-27 15:12:18.971043] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:50.997 [2024-11-27 15:12:19.014247] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:50.997 [2024-11-27 15:12:19.014364] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:51.569 15:12:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:51.569 15:12:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:07:51.569 15:12:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:51.569 15:12:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:51.569 15:12:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.569 [2024-11-27 15:12:19.601505] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:51.569 [2024-11-27 15:12:19.601569] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:51.569 [2024-11-27 15:12:19.601585] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:51.569 [2024-11-27 15:12:19.601598] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:51.569 15:12:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:51.569 15:12:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:51.569 15:12:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:51.569 15:12:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:51.569 15:12:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:51.569 15:12:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:51.569 15:12:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:51.569 15:12:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:51.569 15:12:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:51.569 15:12:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:51.569 15:12:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:51.569 15:12:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:51.569 15:12:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:51.569 15:12:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.569 15:12:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:51.569 15:12:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:51.569 15:12:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:51.569 "name": "Existed_Raid", 00:07:51.569 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:51.569 "strip_size_kb": 0, 00:07:51.569 "state": "configuring", 00:07:51.569 "raid_level": "raid1", 00:07:51.569 "superblock": false, 00:07:51.569 "num_base_bdevs": 2, 00:07:51.569 "num_base_bdevs_discovered": 0, 00:07:51.569 "num_base_bdevs_operational": 2, 00:07:51.569 "base_bdevs_list": [ 00:07:51.569 { 00:07:51.569 "name": "BaseBdev1", 00:07:51.569 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:51.569 "is_configured": false, 00:07:51.569 "data_offset": 0, 00:07:51.569 "data_size": 0 00:07:51.569 }, 00:07:51.569 { 00:07:51.569 "name": "BaseBdev2", 00:07:51.569 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:51.569 "is_configured": false, 00:07:51.569 "data_offset": 0, 00:07:51.569 "data_size": 0 00:07:51.569 } 00:07:51.569 ] 00:07:51.569 }' 00:07:51.569 15:12:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:51.569 15:12:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.138 [2024-11-27 15:12:20.072621] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:52.138 [2024-11-27 15:12:20.072729] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.138 [2024-11-27 15:12:20.084595] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:52.138 [2024-11-27 15:12:20.084683] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:52.138 [2024-11-27 15:12:20.084709] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:52.138 [2024-11-27 15:12:20.084730] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.138 [2024-11-27 15:12:20.105656] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:52.138 BaseBdev1 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.138 [ 00:07:52.138 { 00:07:52.138 "name": "BaseBdev1", 00:07:52.138 "aliases": [ 00:07:52.138 "a9560791-37ad-4e72-bfdf-b8919bb60c30" 00:07:52.138 ], 00:07:52.138 "product_name": "Malloc disk", 00:07:52.138 "block_size": 512, 00:07:52.138 "num_blocks": 65536, 00:07:52.138 "uuid": "a9560791-37ad-4e72-bfdf-b8919bb60c30", 00:07:52.138 "assigned_rate_limits": { 00:07:52.138 "rw_ios_per_sec": 0, 00:07:52.138 "rw_mbytes_per_sec": 0, 00:07:52.138 "r_mbytes_per_sec": 0, 00:07:52.138 "w_mbytes_per_sec": 0 00:07:52.138 }, 00:07:52.138 "claimed": true, 00:07:52.138 "claim_type": "exclusive_write", 00:07:52.138 "zoned": false, 00:07:52.138 "supported_io_types": { 00:07:52.138 "read": true, 00:07:52.138 "write": true, 00:07:52.138 "unmap": true, 00:07:52.138 "flush": true, 00:07:52.138 "reset": true, 00:07:52.138 "nvme_admin": false, 00:07:52.138 "nvme_io": false, 00:07:52.138 "nvme_io_md": false, 00:07:52.138 "write_zeroes": true, 00:07:52.138 "zcopy": true, 00:07:52.138 "get_zone_info": false, 00:07:52.138 "zone_management": false, 00:07:52.138 "zone_append": false, 00:07:52.138 "compare": false, 00:07:52.138 "compare_and_write": false, 00:07:52.138 "abort": true, 00:07:52.138 "seek_hole": false, 00:07:52.138 "seek_data": false, 00:07:52.138 "copy": true, 00:07:52.138 "nvme_iov_md": false 00:07:52.138 }, 00:07:52.138 "memory_domains": [ 00:07:52.138 { 00:07:52.138 "dma_device_id": "system", 00:07:52.138 "dma_device_type": 1 00:07:52.138 }, 00:07:52.138 { 00:07:52.138 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:52.138 "dma_device_type": 2 00:07:52.138 } 00:07:52.138 ], 00:07:52.138 "driver_specific": {} 00:07:52.138 } 00:07:52.138 ] 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:52.138 "name": "Existed_Raid", 00:07:52.138 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:52.138 "strip_size_kb": 0, 00:07:52.138 "state": "configuring", 00:07:52.138 "raid_level": "raid1", 00:07:52.138 "superblock": false, 00:07:52.138 "num_base_bdevs": 2, 00:07:52.138 "num_base_bdevs_discovered": 1, 00:07:52.138 "num_base_bdevs_operational": 2, 00:07:52.138 "base_bdevs_list": [ 00:07:52.138 { 00:07:52.138 "name": "BaseBdev1", 00:07:52.138 "uuid": "a9560791-37ad-4e72-bfdf-b8919bb60c30", 00:07:52.138 "is_configured": true, 00:07:52.138 "data_offset": 0, 00:07:52.138 "data_size": 65536 00:07:52.138 }, 00:07:52.138 { 00:07:52.138 "name": "BaseBdev2", 00:07:52.138 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:52.138 "is_configured": false, 00:07:52.138 "data_offset": 0, 00:07:52.138 "data_size": 0 00:07:52.138 } 00:07:52.138 ] 00:07:52.138 }' 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:52.138 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.707 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:52.707 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.707 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.707 [2024-11-27 15:12:20.552965] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:52.707 [2024-11-27 15:12:20.553020] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:07:52.707 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.707 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:52.707 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.707 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.707 [2024-11-27 15:12:20.564943] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:52.707 [2024-11-27 15:12:20.566715] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:52.707 [2024-11-27 15:12:20.566756] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:52.707 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.707 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:52.707 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:52.707 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:52.707 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:52.707 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:52.707 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:52.707 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:52.707 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:52.707 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:52.707 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:52.707 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:52.707 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:52.707 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:52.707 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.707 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.707 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:52.707 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.707 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:52.707 "name": "Existed_Raid", 00:07:52.707 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:52.707 "strip_size_kb": 0, 00:07:52.707 "state": "configuring", 00:07:52.707 "raid_level": "raid1", 00:07:52.707 "superblock": false, 00:07:52.707 "num_base_bdevs": 2, 00:07:52.707 "num_base_bdevs_discovered": 1, 00:07:52.707 "num_base_bdevs_operational": 2, 00:07:52.707 "base_bdevs_list": [ 00:07:52.707 { 00:07:52.707 "name": "BaseBdev1", 00:07:52.707 "uuid": "a9560791-37ad-4e72-bfdf-b8919bb60c30", 00:07:52.707 "is_configured": true, 00:07:52.707 "data_offset": 0, 00:07:52.707 "data_size": 65536 00:07:52.707 }, 00:07:52.707 { 00:07:52.707 "name": "BaseBdev2", 00:07:52.707 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:52.707 "is_configured": false, 00:07:52.707 "data_offset": 0, 00:07:52.707 "data_size": 0 00:07:52.707 } 00:07:52.707 ] 00:07:52.707 }' 00:07:52.707 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:52.707 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.967 [2024-11-27 15:12:20.931466] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:52.967 [2024-11-27 15:12:20.931529] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:07:52.967 [2024-11-27 15:12:20.931544] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:07:52.967 [2024-11-27 15:12:20.931839] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:07:52.967 [2024-11-27 15:12:20.932037] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:07:52.967 [2024-11-27 15:12:20.932063] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:07:52.967 [2024-11-27 15:12:20.932293] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:52.967 BaseBdev2 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.967 [ 00:07:52.967 { 00:07:52.967 "name": "BaseBdev2", 00:07:52.967 "aliases": [ 00:07:52.967 "6dfc583a-d73d-4f13-9bad-4dfba8b9bb1a" 00:07:52.967 ], 00:07:52.967 "product_name": "Malloc disk", 00:07:52.967 "block_size": 512, 00:07:52.967 "num_blocks": 65536, 00:07:52.967 "uuid": "6dfc583a-d73d-4f13-9bad-4dfba8b9bb1a", 00:07:52.967 "assigned_rate_limits": { 00:07:52.967 "rw_ios_per_sec": 0, 00:07:52.967 "rw_mbytes_per_sec": 0, 00:07:52.967 "r_mbytes_per_sec": 0, 00:07:52.967 "w_mbytes_per_sec": 0 00:07:52.967 }, 00:07:52.967 "claimed": true, 00:07:52.967 "claim_type": "exclusive_write", 00:07:52.967 "zoned": false, 00:07:52.967 "supported_io_types": { 00:07:52.967 "read": true, 00:07:52.967 "write": true, 00:07:52.967 "unmap": true, 00:07:52.967 "flush": true, 00:07:52.967 "reset": true, 00:07:52.967 "nvme_admin": false, 00:07:52.967 "nvme_io": false, 00:07:52.967 "nvme_io_md": false, 00:07:52.967 "write_zeroes": true, 00:07:52.967 "zcopy": true, 00:07:52.967 "get_zone_info": false, 00:07:52.967 "zone_management": false, 00:07:52.967 "zone_append": false, 00:07:52.967 "compare": false, 00:07:52.967 "compare_and_write": false, 00:07:52.967 "abort": true, 00:07:52.967 "seek_hole": false, 00:07:52.967 "seek_data": false, 00:07:52.967 "copy": true, 00:07:52.967 "nvme_iov_md": false 00:07:52.967 }, 00:07:52.967 "memory_domains": [ 00:07:52.967 { 00:07:52.967 "dma_device_id": "system", 00:07:52.967 "dma_device_type": 1 00:07:52.967 }, 00:07:52.967 { 00:07:52.967 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:52.967 "dma_device_type": 2 00:07:52.967 } 00:07:52.967 ], 00:07:52.967 "driver_specific": {} 00:07:52.967 } 00:07:52.967 ] 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.967 15:12:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.967 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:52.967 "name": "Existed_Raid", 00:07:52.967 "uuid": "80f209bf-ae32-4e65-904b-c8ace4841e3c", 00:07:52.967 "strip_size_kb": 0, 00:07:52.967 "state": "online", 00:07:52.967 "raid_level": "raid1", 00:07:52.967 "superblock": false, 00:07:52.967 "num_base_bdevs": 2, 00:07:52.967 "num_base_bdevs_discovered": 2, 00:07:52.967 "num_base_bdevs_operational": 2, 00:07:52.967 "base_bdevs_list": [ 00:07:52.967 { 00:07:52.967 "name": "BaseBdev1", 00:07:52.967 "uuid": "a9560791-37ad-4e72-bfdf-b8919bb60c30", 00:07:52.967 "is_configured": true, 00:07:52.967 "data_offset": 0, 00:07:52.967 "data_size": 65536 00:07:52.967 }, 00:07:52.967 { 00:07:52.967 "name": "BaseBdev2", 00:07:52.967 "uuid": "6dfc583a-d73d-4f13-9bad-4dfba8b9bb1a", 00:07:52.967 "is_configured": true, 00:07:52.967 "data_offset": 0, 00:07:52.967 "data_size": 65536 00:07:52.967 } 00:07:52.967 ] 00:07:52.967 }' 00:07:52.967 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:52.968 15:12:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.536 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:53.536 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:53.536 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:53.536 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:53.536 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:53.536 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:53.536 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:53.536 15:12:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:53.536 15:12:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.536 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:53.536 [2024-11-27 15:12:21.383037] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:53.536 15:12:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:53.536 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:53.536 "name": "Existed_Raid", 00:07:53.536 "aliases": [ 00:07:53.536 "80f209bf-ae32-4e65-904b-c8ace4841e3c" 00:07:53.536 ], 00:07:53.536 "product_name": "Raid Volume", 00:07:53.536 "block_size": 512, 00:07:53.536 "num_blocks": 65536, 00:07:53.536 "uuid": "80f209bf-ae32-4e65-904b-c8ace4841e3c", 00:07:53.536 "assigned_rate_limits": { 00:07:53.536 "rw_ios_per_sec": 0, 00:07:53.536 "rw_mbytes_per_sec": 0, 00:07:53.537 "r_mbytes_per_sec": 0, 00:07:53.537 "w_mbytes_per_sec": 0 00:07:53.537 }, 00:07:53.537 "claimed": false, 00:07:53.537 "zoned": false, 00:07:53.537 "supported_io_types": { 00:07:53.537 "read": true, 00:07:53.537 "write": true, 00:07:53.537 "unmap": false, 00:07:53.537 "flush": false, 00:07:53.537 "reset": true, 00:07:53.537 "nvme_admin": false, 00:07:53.537 "nvme_io": false, 00:07:53.537 "nvme_io_md": false, 00:07:53.537 "write_zeroes": true, 00:07:53.537 "zcopy": false, 00:07:53.537 "get_zone_info": false, 00:07:53.537 "zone_management": false, 00:07:53.537 "zone_append": false, 00:07:53.537 "compare": false, 00:07:53.537 "compare_and_write": false, 00:07:53.537 "abort": false, 00:07:53.537 "seek_hole": false, 00:07:53.537 "seek_data": false, 00:07:53.537 "copy": false, 00:07:53.537 "nvme_iov_md": false 00:07:53.537 }, 00:07:53.537 "memory_domains": [ 00:07:53.537 { 00:07:53.537 "dma_device_id": "system", 00:07:53.537 "dma_device_type": 1 00:07:53.537 }, 00:07:53.537 { 00:07:53.537 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:53.537 "dma_device_type": 2 00:07:53.537 }, 00:07:53.537 { 00:07:53.537 "dma_device_id": "system", 00:07:53.537 "dma_device_type": 1 00:07:53.537 }, 00:07:53.537 { 00:07:53.537 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:53.537 "dma_device_type": 2 00:07:53.537 } 00:07:53.537 ], 00:07:53.537 "driver_specific": { 00:07:53.537 "raid": { 00:07:53.537 "uuid": "80f209bf-ae32-4e65-904b-c8ace4841e3c", 00:07:53.537 "strip_size_kb": 0, 00:07:53.537 "state": "online", 00:07:53.537 "raid_level": "raid1", 00:07:53.537 "superblock": false, 00:07:53.537 "num_base_bdevs": 2, 00:07:53.537 "num_base_bdevs_discovered": 2, 00:07:53.537 "num_base_bdevs_operational": 2, 00:07:53.537 "base_bdevs_list": [ 00:07:53.537 { 00:07:53.537 "name": "BaseBdev1", 00:07:53.537 "uuid": "a9560791-37ad-4e72-bfdf-b8919bb60c30", 00:07:53.537 "is_configured": true, 00:07:53.537 "data_offset": 0, 00:07:53.537 "data_size": 65536 00:07:53.537 }, 00:07:53.537 { 00:07:53.537 "name": "BaseBdev2", 00:07:53.537 "uuid": "6dfc583a-d73d-4f13-9bad-4dfba8b9bb1a", 00:07:53.537 "is_configured": true, 00:07:53.537 "data_offset": 0, 00:07:53.537 "data_size": 65536 00:07:53.537 } 00:07:53.537 ] 00:07:53.537 } 00:07:53.537 } 00:07:53.537 }' 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:53.537 BaseBdev2' 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.537 [2024-11-27 15:12:21.606395] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:53.537 15:12:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.797 15:12:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:53.797 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:53.797 "name": "Existed_Raid", 00:07:53.797 "uuid": "80f209bf-ae32-4e65-904b-c8ace4841e3c", 00:07:53.797 "strip_size_kb": 0, 00:07:53.797 "state": "online", 00:07:53.797 "raid_level": "raid1", 00:07:53.797 "superblock": false, 00:07:53.797 "num_base_bdevs": 2, 00:07:53.797 "num_base_bdevs_discovered": 1, 00:07:53.797 "num_base_bdevs_operational": 1, 00:07:53.797 "base_bdevs_list": [ 00:07:53.797 { 00:07:53.797 "name": null, 00:07:53.797 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:53.797 "is_configured": false, 00:07:53.797 "data_offset": 0, 00:07:53.797 "data_size": 65536 00:07:53.797 }, 00:07:53.797 { 00:07:53.797 "name": "BaseBdev2", 00:07:53.797 "uuid": "6dfc583a-d73d-4f13-9bad-4dfba8b9bb1a", 00:07:53.797 "is_configured": true, 00:07:53.797 "data_offset": 0, 00:07:53.797 "data_size": 65536 00:07:53.797 } 00:07:53.797 ] 00:07:53.797 }' 00:07:53.797 15:12:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:53.797 15:12:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.057 15:12:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:54.057 15:12:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:54.057 15:12:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:54.057 15:12:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:54.057 15:12:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:54.057 15:12:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.057 15:12:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:54.057 15:12:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:54.057 15:12:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:54.057 15:12:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:54.057 15:12:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:54.057 15:12:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.057 [2024-11-27 15:12:22.056848] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:54.057 [2024-11-27 15:12:22.056960] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:54.057 [2024-11-27 15:12:22.068593] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:54.057 [2024-11-27 15:12:22.068645] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:54.057 [2024-11-27 15:12:22.068657] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:07:54.057 15:12:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:54.057 15:12:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:54.057 15:12:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:54.057 15:12:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:54.057 15:12:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:54.057 15:12:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:54.057 15:12:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.057 15:12:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:54.057 15:12:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:54.057 15:12:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:54.057 15:12:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:54.057 15:12:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 74142 00:07:54.057 15:12:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 74142 ']' 00:07:54.057 15:12:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 74142 00:07:54.057 15:12:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:07:54.057 15:12:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:54.057 15:12:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 74142 00:07:54.316 killing process with pid 74142 00:07:54.316 15:12:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:54.316 15:12:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:54.316 15:12:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 74142' 00:07:54.316 15:12:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 74142 00:07:54.316 [2024-11-27 15:12:22.165571] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:54.316 15:12:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 74142 00:07:54.316 [2024-11-27 15:12:22.166530] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:54.316 15:12:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:07:54.316 00:07:54.316 real 0m3.692s 00:07:54.316 user 0m5.781s 00:07:54.316 sys 0m0.743s 00:07:54.316 15:12:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:54.316 15:12:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.316 ************************************ 00:07:54.316 END TEST raid_state_function_test 00:07:54.316 ************************************ 00:07:54.316 15:12:22 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 2 true 00:07:54.317 15:12:22 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:54.576 15:12:22 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:54.576 15:12:22 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:54.576 ************************************ 00:07:54.576 START TEST raid_state_function_test_sb 00:07:54.576 ************************************ 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 true 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=74379 00:07:54.576 Process raid pid: 74379 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 74379' 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 74379 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 74379 ']' 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:54.576 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:54.576 15:12:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:54.576 [2024-11-27 15:12:22.527507] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:07:54.576 [2024-11-27 15:12:22.527656] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:54.835 [2024-11-27 15:12:22.705013] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:54.835 [2024-11-27 15:12:22.730965] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:54.835 [2024-11-27 15:12:22.774166] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:54.835 [2024-11-27 15:12:22.774212] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:55.404 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:55.404 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:07:55.404 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:55.404 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:55.404 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:55.405 [2024-11-27 15:12:23.353812] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:55.405 [2024-11-27 15:12:23.353870] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:55.405 [2024-11-27 15:12:23.353880] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:55.405 [2024-11-27 15:12:23.353891] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:55.405 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:55.405 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:55.405 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:55.405 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:55.405 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:55.405 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:55.405 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:55.405 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:55.405 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:55.405 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:55.405 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:55.405 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:55.405 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:55.405 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:55.405 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:55.405 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:55.405 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:55.405 "name": "Existed_Raid", 00:07:55.405 "uuid": "ac47e087-3dd1-49e1-988e-f9432aec1fba", 00:07:55.405 "strip_size_kb": 0, 00:07:55.405 "state": "configuring", 00:07:55.405 "raid_level": "raid1", 00:07:55.405 "superblock": true, 00:07:55.405 "num_base_bdevs": 2, 00:07:55.405 "num_base_bdevs_discovered": 0, 00:07:55.405 "num_base_bdevs_operational": 2, 00:07:55.405 "base_bdevs_list": [ 00:07:55.405 { 00:07:55.405 "name": "BaseBdev1", 00:07:55.405 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:55.405 "is_configured": false, 00:07:55.405 "data_offset": 0, 00:07:55.405 "data_size": 0 00:07:55.405 }, 00:07:55.405 { 00:07:55.405 "name": "BaseBdev2", 00:07:55.405 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:55.405 "is_configured": false, 00:07:55.405 "data_offset": 0, 00:07:55.405 "data_size": 0 00:07:55.405 } 00:07:55.405 ] 00:07:55.405 }' 00:07:55.405 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:55.405 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:55.664 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:55.664 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:55.664 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:55.925 [2024-11-27 15:12:23.773027] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:55.926 [2024-11-27 15:12:23.773070] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:55.926 [2024-11-27 15:12:23.784998] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:55.926 [2024-11-27 15:12:23.785037] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:55.926 [2024-11-27 15:12:23.785046] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:55.926 [2024-11-27 15:12:23.785055] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:55.926 [2024-11-27 15:12:23.806073] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:55.926 BaseBdev1 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:55.926 [ 00:07:55.926 { 00:07:55.926 "name": "BaseBdev1", 00:07:55.926 "aliases": [ 00:07:55.926 "ae13d507-e640-4774-a7d2-fc6769170847" 00:07:55.926 ], 00:07:55.926 "product_name": "Malloc disk", 00:07:55.926 "block_size": 512, 00:07:55.926 "num_blocks": 65536, 00:07:55.926 "uuid": "ae13d507-e640-4774-a7d2-fc6769170847", 00:07:55.926 "assigned_rate_limits": { 00:07:55.926 "rw_ios_per_sec": 0, 00:07:55.926 "rw_mbytes_per_sec": 0, 00:07:55.926 "r_mbytes_per_sec": 0, 00:07:55.926 "w_mbytes_per_sec": 0 00:07:55.926 }, 00:07:55.926 "claimed": true, 00:07:55.926 "claim_type": "exclusive_write", 00:07:55.926 "zoned": false, 00:07:55.926 "supported_io_types": { 00:07:55.926 "read": true, 00:07:55.926 "write": true, 00:07:55.926 "unmap": true, 00:07:55.926 "flush": true, 00:07:55.926 "reset": true, 00:07:55.926 "nvme_admin": false, 00:07:55.926 "nvme_io": false, 00:07:55.926 "nvme_io_md": false, 00:07:55.926 "write_zeroes": true, 00:07:55.926 "zcopy": true, 00:07:55.926 "get_zone_info": false, 00:07:55.926 "zone_management": false, 00:07:55.926 "zone_append": false, 00:07:55.926 "compare": false, 00:07:55.926 "compare_and_write": false, 00:07:55.926 "abort": true, 00:07:55.926 "seek_hole": false, 00:07:55.926 "seek_data": false, 00:07:55.926 "copy": true, 00:07:55.926 "nvme_iov_md": false 00:07:55.926 }, 00:07:55.926 "memory_domains": [ 00:07:55.926 { 00:07:55.926 "dma_device_id": "system", 00:07:55.926 "dma_device_type": 1 00:07:55.926 }, 00:07:55.926 { 00:07:55.926 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:55.926 "dma_device_type": 2 00:07:55.926 } 00:07:55.926 ], 00:07:55.926 "driver_specific": {} 00:07:55.926 } 00:07:55.926 ] 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:55.926 "name": "Existed_Raid", 00:07:55.926 "uuid": "b9875822-0d13-4421-a333-d601de8e82a1", 00:07:55.926 "strip_size_kb": 0, 00:07:55.926 "state": "configuring", 00:07:55.926 "raid_level": "raid1", 00:07:55.926 "superblock": true, 00:07:55.926 "num_base_bdevs": 2, 00:07:55.926 "num_base_bdevs_discovered": 1, 00:07:55.926 "num_base_bdevs_operational": 2, 00:07:55.926 "base_bdevs_list": [ 00:07:55.926 { 00:07:55.926 "name": "BaseBdev1", 00:07:55.926 "uuid": "ae13d507-e640-4774-a7d2-fc6769170847", 00:07:55.926 "is_configured": true, 00:07:55.926 "data_offset": 2048, 00:07:55.926 "data_size": 63488 00:07:55.926 }, 00:07:55.926 { 00:07:55.926 "name": "BaseBdev2", 00:07:55.926 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:55.926 "is_configured": false, 00:07:55.926 "data_offset": 0, 00:07:55.926 "data_size": 0 00:07:55.926 } 00:07:55.926 ] 00:07:55.926 }' 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:55.926 15:12:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:56.495 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:56.495 15:12:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:56.495 15:12:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:56.495 [2024-11-27 15:12:24.317261] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:56.495 [2024-11-27 15:12:24.317329] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:07:56.495 15:12:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:56.496 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:56.496 15:12:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:56.496 15:12:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:56.496 [2024-11-27 15:12:24.325259] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:56.496 [2024-11-27 15:12:24.327128] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:56.496 [2024-11-27 15:12:24.327172] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:56.496 15:12:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:56.496 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:56.496 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:56.496 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:56.496 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:56.496 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:56.496 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:56.496 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:56.496 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:56.496 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:56.496 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:56.496 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:56.496 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:56.496 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:56.496 15:12:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:56.496 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:56.496 15:12:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:56.496 15:12:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:56.496 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:56.496 "name": "Existed_Raid", 00:07:56.496 "uuid": "cd1611a6-57e9-4174-925f-a59e73fcae07", 00:07:56.496 "strip_size_kb": 0, 00:07:56.496 "state": "configuring", 00:07:56.496 "raid_level": "raid1", 00:07:56.496 "superblock": true, 00:07:56.496 "num_base_bdevs": 2, 00:07:56.496 "num_base_bdevs_discovered": 1, 00:07:56.496 "num_base_bdevs_operational": 2, 00:07:56.496 "base_bdevs_list": [ 00:07:56.496 { 00:07:56.496 "name": "BaseBdev1", 00:07:56.496 "uuid": "ae13d507-e640-4774-a7d2-fc6769170847", 00:07:56.496 "is_configured": true, 00:07:56.496 "data_offset": 2048, 00:07:56.496 "data_size": 63488 00:07:56.496 }, 00:07:56.496 { 00:07:56.496 "name": "BaseBdev2", 00:07:56.496 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:56.496 "is_configured": false, 00:07:56.496 "data_offset": 0, 00:07:56.496 "data_size": 0 00:07:56.496 } 00:07:56.496 ] 00:07:56.496 }' 00:07:56.496 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:56.496 15:12:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:56.757 [2024-11-27 15:12:24.791620] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:56.757 [2024-11-27 15:12:24.791851] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:07:56.757 [2024-11-27 15:12:24.791879] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:07:56.757 [2024-11-27 15:12:24.792185] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:07:56.757 [2024-11-27 15:12:24.792340] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:07:56.757 [2024-11-27 15:12:24.792362] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:07:56.757 BaseBdev2 00:07:56.757 [2024-11-27 15:12:24.792487] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:56.757 [ 00:07:56.757 { 00:07:56.757 "name": "BaseBdev2", 00:07:56.757 "aliases": [ 00:07:56.757 "dee40092-c788-4c9b-9507-82a7436d0117" 00:07:56.757 ], 00:07:56.757 "product_name": "Malloc disk", 00:07:56.757 "block_size": 512, 00:07:56.757 "num_blocks": 65536, 00:07:56.757 "uuid": "dee40092-c788-4c9b-9507-82a7436d0117", 00:07:56.757 "assigned_rate_limits": { 00:07:56.757 "rw_ios_per_sec": 0, 00:07:56.757 "rw_mbytes_per_sec": 0, 00:07:56.757 "r_mbytes_per_sec": 0, 00:07:56.757 "w_mbytes_per_sec": 0 00:07:56.757 }, 00:07:56.757 "claimed": true, 00:07:56.757 "claim_type": "exclusive_write", 00:07:56.757 "zoned": false, 00:07:56.757 "supported_io_types": { 00:07:56.757 "read": true, 00:07:56.757 "write": true, 00:07:56.757 "unmap": true, 00:07:56.757 "flush": true, 00:07:56.757 "reset": true, 00:07:56.757 "nvme_admin": false, 00:07:56.757 "nvme_io": false, 00:07:56.757 "nvme_io_md": false, 00:07:56.757 "write_zeroes": true, 00:07:56.757 "zcopy": true, 00:07:56.757 "get_zone_info": false, 00:07:56.757 "zone_management": false, 00:07:56.757 "zone_append": false, 00:07:56.757 "compare": false, 00:07:56.757 "compare_and_write": false, 00:07:56.757 "abort": true, 00:07:56.757 "seek_hole": false, 00:07:56.757 "seek_data": false, 00:07:56.757 "copy": true, 00:07:56.757 "nvme_iov_md": false 00:07:56.757 }, 00:07:56.757 "memory_domains": [ 00:07:56.757 { 00:07:56.757 "dma_device_id": "system", 00:07:56.757 "dma_device_type": 1 00:07:56.757 }, 00:07:56.757 { 00:07:56.757 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:56.757 "dma_device_type": 2 00:07:56.757 } 00:07:56.757 ], 00:07:56.757 "driver_specific": {} 00:07:56.757 } 00:07:56.757 ] 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:56.757 15:12:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:57.017 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:57.017 "name": "Existed_Raid", 00:07:57.017 "uuid": "cd1611a6-57e9-4174-925f-a59e73fcae07", 00:07:57.017 "strip_size_kb": 0, 00:07:57.017 "state": "online", 00:07:57.017 "raid_level": "raid1", 00:07:57.017 "superblock": true, 00:07:57.017 "num_base_bdevs": 2, 00:07:57.017 "num_base_bdevs_discovered": 2, 00:07:57.017 "num_base_bdevs_operational": 2, 00:07:57.017 "base_bdevs_list": [ 00:07:57.017 { 00:07:57.017 "name": "BaseBdev1", 00:07:57.017 "uuid": "ae13d507-e640-4774-a7d2-fc6769170847", 00:07:57.017 "is_configured": true, 00:07:57.017 "data_offset": 2048, 00:07:57.017 "data_size": 63488 00:07:57.017 }, 00:07:57.017 { 00:07:57.017 "name": "BaseBdev2", 00:07:57.017 "uuid": "dee40092-c788-4c9b-9507-82a7436d0117", 00:07:57.017 "is_configured": true, 00:07:57.017 "data_offset": 2048, 00:07:57.017 "data_size": 63488 00:07:57.017 } 00:07:57.017 ] 00:07:57.017 }' 00:07:57.017 15:12:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:57.017 15:12:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:57.276 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:57.276 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:57.276 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:57.277 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:57.277 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:07:57.277 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:57.277 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:57.277 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:57.277 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:57.277 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:57.277 [2024-11-27 15:12:25.251221] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:57.277 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:57.277 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:57.277 "name": "Existed_Raid", 00:07:57.277 "aliases": [ 00:07:57.277 "cd1611a6-57e9-4174-925f-a59e73fcae07" 00:07:57.277 ], 00:07:57.277 "product_name": "Raid Volume", 00:07:57.277 "block_size": 512, 00:07:57.277 "num_blocks": 63488, 00:07:57.277 "uuid": "cd1611a6-57e9-4174-925f-a59e73fcae07", 00:07:57.277 "assigned_rate_limits": { 00:07:57.277 "rw_ios_per_sec": 0, 00:07:57.277 "rw_mbytes_per_sec": 0, 00:07:57.277 "r_mbytes_per_sec": 0, 00:07:57.277 "w_mbytes_per_sec": 0 00:07:57.277 }, 00:07:57.277 "claimed": false, 00:07:57.277 "zoned": false, 00:07:57.277 "supported_io_types": { 00:07:57.277 "read": true, 00:07:57.277 "write": true, 00:07:57.277 "unmap": false, 00:07:57.277 "flush": false, 00:07:57.277 "reset": true, 00:07:57.277 "nvme_admin": false, 00:07:57.277 "nvme_io": false, 00:07:57.277 "nvme_io_md": false, 00:07:57.277 "write_zeroes": true, 00:07:57.277 "zcopy": false, 00:07:57.277 "get_zone_info": false, 00:07:57.277 "zone_management": false, 00:07:57.277 "zone_append": false, 00:07:57.277 "compare": false, 00:07:57.277 "compare_and_write": false, 00:07:57.277 "abort": false, 00:07:57.277 "seek_hole": false, 00:07:57.277 "seek_data": false, 00:07:57.277 "copy": false, 00:07:57.277 "nvme_iov_md": false 00:07:57.277 }, 00:07:57.277 "memory_domains": [ 00:07:57.277 { 00:07:57.277 "dma_device_id": "system", 00:07:57.277 "dma_device_type": 1 00:07:57.277 }, 00:07:57.277 { 00:07:57.277 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:57.277 "dma_device_type": 2 00:07:57.277 }, 00:07:57.277 { 00:07:57.277 "dma_device_id": "system", 00:07:57.277 "dma_device_type": 1 00:07:57.277 }, 00:07:57.277 { 00:07:57.277 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:57.277 "dma_device_type": 2 00:07:57.277 } 00:07:57.277 ], 00:07:57.277 "driver_specific": { 00:07:57.277 "raid": { 00:07:57.277 "uuid": "cd1611a6-57e9-4174-925f-a59e73fcae07", 00:07:57.277 "strip_size_kb": 0, 00:07:57.277 "state": "online", 00:07:57.277 "raid_level": "raid1", 00:07:57.277 "superblock": true, 00:07:57.277 "num_base_bdevs": 2, 00:07:57.277 "num_base_bdevs_discovered": 2, 00:07:57.277 "num_base_bdevs_operational": 2, 00:07:57.277 "base_bdevs_list": [ 00:07:57.277 { 00:07:57.277 "name": "BaseBdev1", 00:07:57.277 "uuid": "ae13d507-e640-4774-a7d2-fc6769170847", 00:07:57.277 "is_configured": true, 00:07:57.277 "data_offset": 2048, 00:07:57.277 "data_size": 63488 00:07:57.277 }, 00:07:57.277 { 00:07:57.277 "name": "BaseBdev2", 00:07:57.277 "uuid": "dee40092-c788-4c9b-9507-82a7436d0117", 00:07:57.277 "is_configured": true, 00:07:57.277 "data_offset": 2048, 00:07:57.277 "data_size": 63488 00:07:57.277 } 00:07:57.277 ] 00:07:57.277 } 00:07:57.277 } 00:07:57.277 }' 00:07:57.277 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:57.277 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:57.277 BaseBdev2' 00:07:57.277 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:57.277 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:57.277 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:57.277 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:57.277 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:57.277 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:57.277 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:57.537 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:57.537 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:57.538 [2024-11-27 15:12:25.454574] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:57.538 "name": "Existed_Raid", 00:07:57.538 "uuid": "cd1611a6-57e9-4174-925f-a59e73fcae07", 00:07:57.538 "strip_size_kb": 0, 00:07:57.538 "state": "online", 00:07:57.538 "raid_level": "raid1", 00:07:57.538 "superblock": true, 00:07:57.538 "num_base_bdevs": 2, 00:07:57.538 "num_base_bdevs_discovered": 1, 00:07:57.538 "num_base_bdevs_operational": 1, 00:07:57.538 "base_bdevs_list": [ 00:07:57.538 { 00:07:57.538 "name": null, 00:07:57.538 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:57.538 "is_configured": false, 00:07:57.538 "data_offset": 0, 00:07:57.538 "data_size": 63488 00:07:57.538 }, 00:07:57.538 { 00:07:57.538 "name": "BaseBdev2", 00:07:57.538 "uuid": "dee40092-c788-4c9b-9507-82a7436d0117", 00:07:57.538 "is_configured": true, 00:07:57.538 "data_offset": 2048, 00:07:57.538 "data_size": 63488 00:07:57.538 } 00:07:57.538 ] 00:07:57.538 }' 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:57.538 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:57.798 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:57.798 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:57.798 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:57.798 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:57.798 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:57.798 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:57.798 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:58.058 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:58.058 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:58.058 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:58.058 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:58.058 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.058 [2024-11-27 15:12:25.917420] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:58.058 [2024-11-27 15:12:25.917540] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:58.058 [2024-11-27 15:12:25.929288] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:58.058 [2024-11-27 15:12:25.929340] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:58.058 [2024-11-27 15:12:25.929352] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:07:58.058 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:58.058 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:58.058 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:58.058 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:58.058 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:58.058 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:58.058 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.058 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:58.058 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:58.058 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:58.058 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:58.058 15:12:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 74379 00:07:58.058 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 74379 ']' 00:07:58.058 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 74379 00:07:58.058 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:07:58.058 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:58.058 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 74379 00:07:58.058 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:58.058 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:58.058 killing process with pid 74379 00:07:58.058 15:12:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 74379' 00:07:58.058 15:12:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 74379 00:07:58.058 [2024-11-27 15:12:26.001394] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:58.058 15:12:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 74379 00:07:58.058 [2024-11-27 15:12:26.002392] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:58.318 15:12:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:07:58.318 00:07:58.318 real 0m3.792s 00:07:58.318 user 0m5.968s 00:07:58.318 sys 0m0.788s 00:07:58.318 15:12:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:58.318 15:12:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.318 ************************************ 00:07:58.318 END TEST raid_state_function_test_sb 00:07:58.318 ************************************ 00:07:58.318 15:12:26 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 2 00:07:58.318 15:12:26 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:07:58.318 15:12:26 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:58.318 15:12:26 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:58.318 ************************************ 00:07:58.318 START TEST raid_superblock_test 00:07:58.318 ************************************ 00:07:58.318 15:12:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 2 00:07:58.318 15:12:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:07:58.318 15:12:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:07:58.318 15:12:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:07:58.318 15:12:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:07:58.318 15:12:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:07:58.318 15:12:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:07:58.318 15:12:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:07:58.318 15:12:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:07:58.318 15:12:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:07:58.318 15:12:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:07:58.318 15:12:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:07:58.318 15:12:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:07:58.318 15:12:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:07:58.318 15:12:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:07:58.318 15:12:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:07:58.318 15:12:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=74620 00:07:58.318 15:12:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:07:58.318 15:12:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 74620 00:07:58.318 15:12:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 74620 ']' 00:07:58.318 15:12:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:58.318 15:12:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:58.318 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:58.318 15:12:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:58.318 15:12:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:58.318 15:12:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:58.318 [2024-11-27 15:12:26.380758] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:07:58.319 [2024-11-27 15:12:26.380893] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid74620 ] 00:07:58.578 [2024-11-27 15:12:26.550054] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:58.578 [2024-11-27 15:12:26.579319] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:58.578 [2024-11-27 15:12:26.623681] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:58.578 [2024-11-27 15:12:26.623722] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:59.146 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:59.146 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:07:59.146 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:07:59.146 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:59.146 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:07:59.146 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:07:59.146 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:07:59.146 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:59.146 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:59.146 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:59.146 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:07:59.146 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.146 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.146 malloc1 00:07:59.146 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.146 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:59.146 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.146 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.146 [2024-11-27 15:12:27.248924] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:59.146 [2024-11-27 15:12:27.249001] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:59.146 [2024-11-27 15:12:27.249023] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:07:59.146 [2024-11-27 15:12:27.249036] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:59.425 [2024-11-27 15:12:27.251239] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:59.425 [2024-11-27 15:12:27.251283] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:59.425 pt1 00:07:59.425 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.425 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:59.425 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:59.425 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:07:59.425 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:07:59.425 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:07:59.425 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:59.425 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:59.425 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:59.425 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:07:59.425 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.425 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.425 malloc2 00:07:59.425 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.425 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:59.425 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.425 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.425 [2024-11-27 15:12:27.278078] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:59.425 [2024-11-27 15:12:27.278150] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:59.426 [2024-11-27 15:12:27.278170] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:07:59.426 [2024-11-27 15:12:27.278180] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:59.426 [2024-11-27 15:12:27.280461] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:59.426 [2024-11-27 15:12:27.280516] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:59.426 pt2 00:07:59.426 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.426 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:59.426 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:59.426 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:07:59.426 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.426 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.426 [2024-11-27 15:12:27.290108] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:59.426 [2024-11-27 15:12:27.292099] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:59.426 [2024-11-27 15:12:27.292257] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:07:59.426 [2024-11-27 15:12:27.292278] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:07:59.426 [2024-11-27 15:12:27.292575] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:07:59.426 [2024-11-27 15:12:27.292738] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:07:59.426 [2024-11-27 15:12:27.292754] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:07:59.426 [2024-11-27 15:12:27.292918] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:59.426 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.426 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:07:59.426 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:59.426 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:59.426 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:59.426 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:59.426 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:59.426 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:59.426 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:59.426 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:59.426 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:59.426 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:59.426 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:59.426 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.426 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.426 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.426 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:59.426 "name": "raid_bdev1", 00:07:59.426 "uuid": "dec1cbb5-db8b-4164-88b6-688e44fd41c8", 00:07:59.426 "strip_size_kb": 0, 00:07:59.426 "state": "online", 00:07:59.426 "raid_level": "raid1", 00:07:59.426 "superblock": true, 00:07:59.426 "num_base_bdevs": 2, 00:07:59.426 "num_base_bdevs_discovered": 2, 00:07:59.426 "num_base_bdevs_operational": 2, 00:07:59.426 "base_bdevs_list": [ 00:07:59.426 { 00:07:59.426 "name": "pt1", 00:07:59.426 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:59.426 "is_configured": true, 00:07:59.426 "data_offset": 2048, 00:07:59.426 "data_size": 63488 00:07:59.426 }, 00:07:59.426 { 00:07:59.426 "name": "pt2", 00:07:59.426 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:59.426 "is_configured": true, 00:07:59.426 "data_offset": 2048, 00:07:59.426 "data_size": 63488 00:07:59.426 } 00:07:59.426 ] 00:07:59.426 }' 00:07:59.426 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:59.426 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.689 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:07:59.689 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:59.689 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:59.689 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:59.689 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:59.689 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:59.689 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:59.689 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.689 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.689 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:59.689 [2024-11-27 15:12:27.709696] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:59.689 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.689 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:59.689 "name": "raid_bdev1", 00:07:59.689 "aliases": [ 00:07:59.689 "dec1cbb5-db8b-4164-88b6-688e44fd41c8" 00:07:59.689 ], 00:07:59.689 "product_name": "Raid Volume", 00:07:59.689 "block_size": 512, 00:07:59.689 "num_blocks": 63488, 00:07:59.689 "uuid": "dec1cbb5-db8b-4164-88b6-688e44fd41c8", 00:07:59.689 "assigned_rate_limits": { 00:07:59.689 "rw_ios_per_sec": 0, 00:07:59.689 "rw_mbytes_per_sec": 0, 00:07:59.689 "r_mbytes_per_sec": 0, 00:07:59.689 "w_mbytes_per_sec": 0 00:07:59.689 }, 00:07:59.689 "claimed": false, 00:07:59.689 "zoned": false, 00:07:59.689 "supported_io_types": { 00:07:59.689 "read": true, 00:07:59.689 "write": true, 00:07:59.689 "unmap": false, 00:07:59.689 "flush": false, 00:07:59.689 "reset": true, 00:07:59.689 "nvme_admin": false, 00:07:59.689 "nvme_io": false, 00:07:59.689 "nvme_io_md": false, 00:07:59.689 "write_zeroes": true, 00:07:59.689 "zcopy": false, 00:07:59.689 "get_zone_info": false, 00:07:59.689 "zone_management": false, 00:07:59.689 "zone_append": false, 00:07:59.689 "compare": false, 00:07:59.689 "compare_and_write": false, 00:07:59.689 "abort": false, 00:07:59.689 "seek_hole": false, 00:07:59.689 "seek_data": false, 00:07:59.689 "copy": false, 00:07:59.689 "nvme_iov_md": false 00:07:59.689 }, 00:07:59.689 "memory_domains": [ 00:07:59.689 { 00:07:59.689 "dma_device_id": "system", 00:07:59.689 "dma_device_type": 1 00:07:59.689 }, 00:07:59.689 { 00:07:59.689 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:59.689 "dma_device_type": 2 00:07:59.689 }, 00:07:59.689 { 00:07:59.689 "dma_device_id": "system", 00:07:59.689 "dma_device_type": 1 00:07:59.689 }, 00:07:59.689 { 00:07:59.689 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:59.689 "dma_device_type": 2 00:07:59.689 } 00:07:59.689 ], 00:07:59.689 "driver_specific": { 00:07:59.689 "raid": { 00:07:59.689 "uuid": "dec1cbb5-db8b-4164-88b6-688e44fd41c8", 00:07:59.689 "strip_size_kb": 0, 00:07:59.689 "state": "online", 00:07:59.689 "raid_level": "raid1", 00:07:59.689 "superblock": true, 00:07:59.689 "num_base_bdevs": 2, 00:07:59.689 "num_base_bdevs_discovered": 2, 00:07:59.689 "num_base_bdevs_operational": 2, 00:07:59.689 "base_bdevs_list": [ 00:07:59.689 { 00:07:59.689 "name": "pt1", 00:07:59.689 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:59.689 "is_configured": true, 00:07:59.689 "data_offset": 2048, 00:07:59.689 "data_size": 63488 00:07:59.689 }, 00:07:59.689 { 00:07:59.689 "name": "pt2", 00:07:59.689 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:59.689 "is_configured": true, 00:07:59.689 "data_offset": 2048, 00:07:59.689 "data_size": 63488 00:07:59.689 } 00:07:59.689 ] 00:07:59.689 } 00:07:59.689 } 00:07:59.689 }' 00:07:59.689 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:59.948 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:59.948 pt2' 00:07:59.948 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:59.948 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:59.948 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:59.948 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:59.948 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:59.948 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.948 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.948 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.948 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:59.948 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:59.948 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:59.948 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:59.948 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:59.948 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.948 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.948 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.948 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:59.948 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:59.948 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:59.949 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:07:59.949 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.949 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.949 [2024-11-27 15:12:27.957291] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:59.949 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.949 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=dec1cbb5-db8b-4164-88b6-688e44fd41c8 00:07:59.949 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z dec1cbb5-db8b-4164-88b6-688e44fd41c8 ']' 00:07:59.949 15:12:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:59.949 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.949 15:12:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.949 [2024-11-27 15:12:28.004943] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:59.949 [2024-11-27 15:12:28.004983] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:59.949 [2024-11-27 15:12:28.005076] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:59.949 [2024-11-27 15:12:28.005141] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:59.949 [2024-11-27 15:12:28.005160] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:07:59.949 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.949 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:59.949 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:07:59.949 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.949 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.949 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.208 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:08:00.208 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:08:00.208 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:00.208 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:08:00.208 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.208 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.208 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.208 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:00.208 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:08:00.208 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.208 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.209 [2024-11-27 15:12:28.140754] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:08:00.209 [2024-11-27 15:12:28.142668] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:08:00.209 [2024-11-27 15:12:28.142743] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:08:00.209 [2024-11-27 15:12:28.142792] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:08:00.209 [2024-11-27 15:12:28.142808] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:00.209 [2024-11-27 15:12:28.142827] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state configuring 00:08:00.209 request: 00:08:00.209 { 00:08:00.209 "name": "raid_bdev1", 00:08:00.209 "raid_level": "raid1", 00:08:00.209 "base_bdevs": [ 00:08:00.209 "malloc1", 00:08:00.209 "malloc2" 00:08:00.209 ], 00:08:00.209 "superblock": false, 00:08:00.209 "method": "bdev_raid_create", 00:08:00.209 "req_id": 1 00:08:00.209 } 00:08:00.209 Got JSON-RPC error response 00:08:00.209 response: 00:08:00.209 { 00:08:00.209 "code": -17, 00:08:00.209 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:08:00.209 } 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.209 [2024-11-27 15:12:28.204570] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:00.209 [2024-11-27 15:12:28.204648] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:00.209 [2024-11-27 15:12:28.204669] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:08:00.209 [2024-11-27 15:12:28.204678] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:00.209 [2024-11-27 15:12:28.206866] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:00.209 [2024-11-27 15:12:28.206920] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:00.209 [2024-11-27 15:12:28.207011] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:08:00.209 [2024-11-27 15:12:28.207051] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:00.209 pt1 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:00.209 "name": "raid_bdev1", 00:08:00.209 "uuid": "dec1cbb5-db8b-4164-88b6-688e44fd41c8", 00:08:00.209 "strip_size_kb": 0, 00:08:00.209 "state": "configuring", 00:08:00.209 "raid_level": "raid1", 00:08:00.209 "superblock": true, 00:08:00.209 "num_base_bdevs": 2, 00:08:00.209 "num_base_bdevs_discovered": 1, 00:08:00.209 "num_base_bdevs_operational": 2, 00:08:00.209 "base_bdevs_list": [ 00:08:00.209 { 00:08:00.209 "name": "pt1", 00:08:00.209 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:00.209 "is_configured": true, 00:08:00.209 "data_offset": 2048, 00:08:00.209 "data_size": 63488 00:08:00.209 }, 00:08:00.209 { 00:08:00.209 "name": null, 00:08:00.209 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:00.209 "is_configured": false, 00:08:00.209 "data_offset": 2048, 00:08:00.209 "data_size": 63488 00:08:00.209 } 00:08:00.209 ] 00:08:00.209 }' 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:00.209 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.778 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:08:00.778 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:08:00.778 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:00.778 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:00.778 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.778 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.778 [2024-11-27 15:12:28.667976] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:00.778 [2024-11-27 15:12:28.668058] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:00.778 [2024-11-27 15:12:28.668083] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:08:00.778 [2024-11-27 15:12:28.668093] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:00.778 [2024-11-27 15:12:28.668525] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:00.778 [2024-11-27 15:12:28.668553] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:00.778 [2024-11-27 15:12:28.668634] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:00.778 [2024-11-27 15:12:28.668656] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:00.778 [2024-11-27 15:12:28.668769] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:08:00.778 [2024-11-27 15:12:28.668790] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:00.778 [2024-11-27 15:12:28.669036] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:08:00.778 [2024-11-27 15:12:28.669161] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:08:00.778 [2024-11-27 15:12:28.669176] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:08:00.778 [2024-11-27 15:12:28.669280] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:00.778 pt2 00:08:00.778 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.778 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:00.778 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:00.778 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:08:00.778 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:00.778 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:00.778 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:00.778 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:00.778 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:00.778 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:00.778 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:00.778 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:00.779 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:00.779 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:00.779 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:00.779 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.779 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.779 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.779 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:00.779 "name": "raid_bdev1", 00:08:00.779 "uuid": "dec1cbb5-db8b-4164-88b6-688e44fd41c8", 00:08:00.779 "strip_size_kb": 0, 00:08:00.779 "state": "online", 00:08:00.779 "raid_level": "raid1", 00:08:00.779 "superblock": true, 00:08:00.779 "num_base_bdevs": 2, 00:08:00.779 "num_base_bdevs_discovered": 2, 00:08:00.779 "num_base_bdevs_operational": 2, 00:08:00.779 "base_bdevs_list": [ 00:08:00.779 { 00:08:00.779 "name": "pt1", 00:08:00.779 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:00.779 "is_configured": true, 00:08:00.779 "data_offset": 2048, 00:08:00.779 "data_size": 63488 00:08:00.779 }, 00:08:00.779 { 00:08:00.779 "name": "pt2", 00:08:00.779 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:00.779 "is_configured": true, 00:08:00.779 "data_offset": 2048, 00:08:00.779 "data_size": 63488 00:08:00.779 } 00:08:00.779 ] 00:08:00.779 }' 00:08:00.779 15:12:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:00.779 15:12:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.039 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:08:01.039 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:01.039 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:01.039 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:01.039 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:01.039 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:01.039 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:01.039 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.039 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.039 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:01.039 [2024-11-27 15:12:29.099558] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:01.039 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.039 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:01.039 "name": "raid_bdev1", 00:08:01.039 "aliases": [ 00:08:01.039 "dec1cbb5-db8b-4164-88b6-688e44fd41c8" 00:08:01.039 ], 00:08:01.039 "product_name": "Raid Volume", 00:08:01.039 "block_size": 512, 00:08:01.039 "num_blocks": 63488, 00:08:01.039 "uuid": "dec1cbb5-db8b-4164-88b6-688e44fd41c8", 00:08:01.039 "assigned_rate_limits": { 00:08:01.039 "rw_ios_per_sec": 0, 00:08:01.039 "rw_mbytes_per_sec": 0, 00:08:01.039 "r_mbytes_per_sec": 0, 00:08:01.039 "w_mbytes_per_sec": 0 00:08:01.039 }, 00:08:01.039 "claimed": false, 00:08:01.039 "zoned": false, 00:08:01.039 "supported_io_types": { 00:08:01.039 "read": true, 00:08:01.039 "write": true, 00:08:01.039 "unmap": false, 00:08:01.039 "flush": false, 00:08:01.039 "reset": true, 00:08:01.039 "nvme_admin": false, 00:08:01.039 "nvme_io": false, 00:08:01.039 "nvme_io_md": false, 00:08:01.039 "write_zeroes": true, 00:08:01.039 "zcopy": false, 00:08:01.039 "get_zone_info": false, 00:08:01.039 "zone_management": false, 00:08:01.039 "zone_append": false, 00:08:01.039 "compare": false, 00:08:01.039 "compare_and_write": false, 00:08:01.039 "abort": false, 00:08:01.039 "seek_hole": false, 00:08:01.039 "seek_data": false, 00:08:01.039 "copy": false, 00:08:01.039 "nvme_iov_md": false 00:08:01.039 }, 00:08:01.039 "memory_domains": [ 00:08:01.039 { 00:08:01.039 "dma_device_id": "system", 00:08:01.039 "dma_device_type": 1 00:08:01.039 }, 00:08:01.039 { 00:08:01.039 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:01.039 "dma_device_type": 2 00:08:01.039 }, 00:08:01.039 { 00:08:01.039 "dma_device_id": "system", 00:08:01.039 "dma_device_type": 1 00:08:01.039 }, 00:08:01.039 { 00:08:01.039 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:01.039 "dma_device_type": 2 00:08:01.039 } 00:08:01.039 ], 00:08:01.039 "driver_specific": { 00:08:01.039 "raid": { 00:08:01.039 "uuid": "dec1cbb5-db8b-4164-88b6-688e44fd41c8", 00:08:01.039 "strip_size_kb": 0, 00:08:01.039 "state": "online", 00:08:01.039 "raid_level": "raid1", 00:08:01.039 "superblock": true, 00:08:01.039 "num_base_bdevs": 2, 00:08:01.039 "num_base_bdevs_discovered": 2, 00:08:01.039 "num_base_bdevs_operational": 2, 00:08:01.039 "base_bdevs_list": [ 00:08:01.039 { 00:08:01.039 "name": "pt1", 00:08:01.039 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:01.039 "is_configured": true, 00:08:01.039 "data_offset": 2048, 00:08:01.039 "data_size": 63488 00:08:01.039 }, 00:08:01.039 { 00:08:01.039 "name": "pt2", 00:08:01.039 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:01.039 "is_configured": true, 00:08:01.039 "data_offset": 2048, 00:08:01.039 "data_size": 63488 00:08:01.039 } 00:08:01.039 ] 00:08:01.039 } 00:08:01.039 } 00:08:01.039 }' 00:08:01.298 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:01.299 pt2' 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:08:01.299 [2024-11-27 15:12:29.303283] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' dec1cbb5-db8b-4164-88b6-688e44fd41c8 '!=' dec1cbb5-db8b-4164-88b6-688e44fd41c8 ']' 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.299 [2024-11-27 15:12:29.350992] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.299 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.557 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:01.557 "name": "raid_bdev1", 00:08:01.557 "uuid": "dec1cbb5-db8b-4164-88b6-688e44fd41c8", 00:08:01.557 "strip_size_kb": 0, 00:08:01.557 "state": "online", 00:08:01.557 "raid_level": "raid1", 00:08:01.557 "superblock": true, 00:08:01.557 "num_base_bdevs": 2, 00:08:01.557 "num_base_bdevs_discovered": 1, 00:08:01.557 "num_base_bdevs_operational": 1, 00:08:01.557 "base_bdevs_list": [ 00:08:01.557 { 00:08:01.557 "name": null, 00:08:01.557 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:01.557 "is_configured": false, 00:08:01.557 "data_offset": 0, 00:08:01.557 "data_size": 63488 00:08:01.557 }, 00:08:01.557 { 00:08:01.557 "name": "pt2", 00:08:01.557 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:01.557 "is_configured": true, 00:08:01.557 "data_offset": 2048, 00:08:01.557 "data_size": 63488 00:08:01.557 } 00:08:01.557 ] 00:08:01.557 }' 00:08:01.557 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:01.557 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.817 [2024-11-27 15:12:29.758201] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:01.817 [2024-11-27 15:12:29.758241] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:01.817 [2024-11-27 15:12:29.758328] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:01.817 [2024-11-27 15:12:29.758378] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:01.817 [2024-11-27 15:12:29.758396] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=1 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.817 [2024-11-27 15:12:29.830089] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:01.817 [2024-11-27 15:12:29.830166] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:01.817 [2024-11-27 15:12:29.830186] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:08:01.817 [2024-11-27 15:12:29.830195] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:01.817 [2024-11-27 15:12:29.832431] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:01.817 [2024-11-27 15:12:29.832479] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:01.817 [2024-11-27 15:12:29.832566] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:01.817 [2024-11-27 15:12:29.832600] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:01.817 [2024-11-27 15:12:29.832679] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:08:01.817 [2024-11-27 15:12:29.832688] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:01.817 [2024-11-27 15:12:29.832923] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:08:01.817 [2024-11-27 15:12:29.833062] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:08:01.817 [2024-11-27 15:12:29.833083] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006d00 00:08:01.817 [2024-11-27 15:12:29.833201] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:01.817 pt2 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:01.817 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:01.818 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:01.818 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:01.818 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.818 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.818 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:01.818 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.818 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:01.818 "name": "raid_bdev1", 00:08:01.818 "uuid": "dec1cbb5-db8b-4164-88b6-688e44fd41c8", 00:08:01.818 "strip_size_kb": 0, 00:08:01.818 "state": "online", 00:08:01.818 "raid_level": "raid1", 00:08:01.818 "superblock": true, 00:08:01.818 "num_base_bdevs": 2, 00:08:01.818 "num_base_bdevs_discovered": 1, 00:08:01.818 "num_base_bdevs_operational": 1, 00:08:01.818 "base_bdevs_list": [ 00:08:01.818 { 00:08:01.818 "name": null, 00:08:01.818 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:01.818 "is_configured": false, 00:08:01.818 "data_offset": 2048, 00:08:01.818 "data_size": 63488 00:08:01.818 }, 00:08:01.818 { 00:08:01.818 "name": "pt2", 00:08:01.818 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:01.818 "is_configured": true, 00:08:01.818 "data_offset": 2048, 00:08:01.818 "data_size": 63488 00:08:01.818 } 00:08:01.818 ] 00:08:01.818 }' 00:08:01.818 15:12:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:01.818 15:12:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.387 15:12:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:02.387 15:12:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:02.387 15:12:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.387 [2024-11-27 15:12:30.329263] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:02.387 [2024-11-27 15:12:30.329303] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:02.387 [2024-11-27 15:12:30.329391] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:02.387 [2024-11-27 15:12:30.329441] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:02.387 [2024-11-27 15:12:30.329454] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name raid_bdev1, state offline 00:08:02.387 15:12:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:02.387 15:12:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:02.387 15:12:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:02.387 15:12:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:08:02.387 15:12:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.387 15:12:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:02.387 15:12:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:08:02.387 15:12:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:08:02.387 15:12:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:08:02.387 15:12:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:02.387 15:12:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:02.387 15:12:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.387 [2024-11-27 15:12:30.389128] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:02.387 [2024-11-27 15:12:30.389210] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:02.387 [2024-11-27 15:12:30.389228] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:08:02.387 [2024-11-27 15:12:30.389241] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:02.387 [2024-11-27 15:12:30.391429] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:02.387 [2024-11-27 15:12:30.391475] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:02.387 [2024-11-27 15:12:30.391561] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:08:02.387 [2024-11-27 15:12:30.391607] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:02.387 [2024-11-27 15:12:30.391720] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:08:02.387 [2024-11-27 15:12:30.391736] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:02.387 [2024-11-27 15:12:30.391766] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007080 name raid_bdev1, state configuring 00:08:02.387 [2024-11-27 15:12:30.391812] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:02.387 [2024-11-27 15:12:30.391939] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007400 00:08:02.387 [2024-11-27 15:12:30.391952] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:02.387 [2024-11-27 15:12:30.392177] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:08:02.387 [2024-11-27 15:12:30.392309] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007400 00:08:02.387 [2024-11-27 15:12:30.392329] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007400 00:08:02.387 [2024-11-27 15:12:30.392445] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:02.387 pt1 00:08:02.387 15:12:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:02.387 15:12:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:08:02.387 15:12:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:08:02.387 15:12:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:02.387 15:12:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:02.387 15:12:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:02.387 15:12:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:02.387 15:12:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:08:02.388 15:12:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:02.388 15:12:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:02.388 15:12:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:02.388 15:12:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:02.388 15:12:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:02.388 15:12:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:02.388 15:12:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:02.388 15:12:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.388 15:12:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:02.388 15:12:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:02.388 "name": "raid_bdev1", 00:08:02.388 "uuid": "dec1cbb5-db8b-4164-88b6-688e44fd41c8", 00:08:02.388 "strip_size_kb": 0, 00:08:02.388 "state": "online", 00:08:02.388 "raid_level": "raid1", 00:08:02.388 "superblock": true, 00:08:02.388 "num_base_bdevs": 2, 00:08:02.388 "num_base_bdevs_discovered": 1, 00:08:02.388 "num_base_bdevs_operational": 1, 00:08:02.388 "base_bdevs_list": [ 00:08:02.388 { 00:08:02.388 "name": null, 00:08:02.388 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:02.388 "is_configured": false, 00:08:02.388 "data_offset": 2048, 00:08:02.388 "data_size": 63488 00:08:02.388 }, 00:08:02.388 { 00:08:02.388 "name": "pt2", 00:08:02.388 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:02.388 "is_configured": true, 00:08:02.388 "data_offset": 2048, 00:08:02.388 "data_size": 63488 00:08:02.388 } 00:08:02.388 ] 00:08:02.388 }' 00:08:02.388 15:12:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:02.388 15:12:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.958 15:12:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:08:02.958 15:12:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:08:02.958 15:12:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:02.958 15:12:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.958 15:12:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:02.958 15:12:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:08:02.958 15:12:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:02.958 15:12:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:08:02.958 15:12:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:02.958 15:12:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.958 [2024-11-27 15:12:30.840622] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:02.958 15:12:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:02.958 15:12:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' dec1cbb5-db8b-4164-88b6-688e44fd41c8 '!=' dec1cbb5-db8b-4164-88b6-688e44fd41c8 ']' 00:08:02.958 15:12:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 74620 00:08:02.958 15:12:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 74620 ']' 00:08:02.958 15:12:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 74620 00:08:02.958 15:12:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:08:02.958 15:12:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:02.958 15:12:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 74620 00:08:02.958 15:12:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:02.958 15:12:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:02.958 killing process with pid 74620 00:08:02.958 15:12:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 74620' 00:08:02.958 15:12:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 74620 00:08:02.958 [2024-11-27 15:12:30.925157] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:02.958 [2024-11-27 15:12:30.925290] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:02.958 15:12:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 74620 00:08:02.958 [2024-11-27 15:12:30.925353] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:02.958 [2024-11-27 15:12:30.925363] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007400 name raid_bdev1, state offline 00:08:02.958 [2024-11-27 15:12:30.949346] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:03.219 15:12:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:08:03.219 00:08:03.219 real 0m4.879s 00:08:03.219 user 0m7.975s 00:08:03.219 sys 0m1.061s 00:08:03.219 15:12:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:03.219 15:12:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:03.219 ************************************ 00:08:03.219 END TEST raid_superblock_test 00:08:03.219 ************************************ 00:08:03.219 15:12:31 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 2 read 00:08:03.219 15:12:31 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:03.219 15:12:31 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:03.219 15:12:31 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:03.219 ************************************ 00:08:03.219 START TEST raid_read_error_test 00:08:03.219 ************************************ 00:08:03.219 15:12:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 2 read 00:08:03.219 15:12:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:08:03.219 15:12:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:08:03.219 15:12:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:08:03.219 15:12:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:03.219 15:12:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:03.219 15:12:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:03.219 15:12:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:03.219 15:12:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:03.219 15:12:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:03.219 15:12:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:03.219 15:12:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:03.219 15:12:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:08:03.219 15:12:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:03.219 15:12:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:03.219 15:12:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:03.219 15:12:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:03.219 15:12:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:03.219 15:12:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:03.219 15:12:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:08:03.219 15:12:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:08:03.219 15:12:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:03.219 15:12:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.2QgjMJ8Fm2 00:08:03.219 15:12:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=74939 00:08:03.219 15:12:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:03.219 15:12:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 74939 00:08:03.219 15:12:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 74939 ']' 00:08:03.219 15:12:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:03.219 15:12:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:03.219 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:03.219 15:12:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:03.219 15:12:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:03.219 15:12:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:03.480 [2024-11-27 15:12:31.360835] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:08:03.480 [2024-11-27 15:12:31.361529] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid74939 ] 00:08:03.480 [2024-11-27 15:12:31.536739] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:03.480 [2024-11-27 15:12:31.566850] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:03.740 [2024-11-27 15:12:31.610670] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:03.740 [2024-11-27 15:12:31.610712] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:04.308 BaseBdev1_malloc 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:04.308 true 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:04.308 [2024-11-27 15:12:32.231278] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:04.308 [2024-11-27 15:12:32.231349] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:04.308 [2024-11-27 15:12:32.231390] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:08:04.308 [2024-11-27 15:12:32.231400] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:04.308 [2024-11-27 15:12:32.233798] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:04.308 [2024-11-27 15:12:32.233840] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:04.308 BaseBdev1 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:04.308 BaseBdev2_malloc 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:04.308 true 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:04.308 [2024-11-27 15:12:32.272427] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:04.308 [2024-11-27 15:12:32.272500] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:04.308 [2024-11-27 15:12:32.272523] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:08:04.308 [2024-11-27 15:12:32.272532] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:04.308 [2024-11-27 15:12:32.274718] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:04.308 [2024-11-27 15:12:32.274762] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:04.308 BaseBdev2 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:04.308 [2024-11-27 15:12:32.284475] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:04.308 [2024-11-27 15:12:32.286387] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:04.308 [2024-11-27 15:12:32.286580] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:08:04.308 [2024-11-27 15:12:32.286593] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:04.308 [2024-11-27 15:12:32.286893] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:08:04.308 [2024-11-27 15:12:32.287084] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:08:04.308 [2024-11-27 15:12:32.287104] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:08:04.308 [2024-11-27 15:12:32.287267] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:04.308 15:12:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:04.309 15:12:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:04.309 15:12:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:04.309 15:12:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:04.309 15:12:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:04.309 15:12:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:04.309 15:12:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:04.309 15:12:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:04.309 15:12:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:04.309 15:12:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:04.309 15:12:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:04.309 15:12:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:04.309 15:12:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:04.309 "name": "raid_bdev1", 00:08:04.309 "uuid": "4b1f6f1a-5bf9-438d-82cc-45e8b27bba8e", 00:08:04.309 "strip_size_kb": 0, 00:08:04.309 "state": "online", 00:08:04.309 "raid_level": "raid1", 00:08:04.309 "superblock": true, 00:08:04.309 "num_base_bdevs": 2, 00:08:04.309 "num_base_bdevs_discovered": 2, 00:08:04.309 "num_base_bdevs_operational": 2, 00:08:04.309 "base_bdevs_list": [ 00:08:04.309 { 00:08:04.309 "name": "BaseBdev1", 00:08:04.309 "uuid": "8f27001d-e1a4-5694-b082-5cb006432068", 00:08:04.309 "is_configured": true, 00:08:04.309 "data_offset": 2048, 00:08:04.309 "data_size": 63488 00:08:04.309 }, 00:08:04.309 { 00:08:04.309 "name": "BaseBdev2", 00:08:04.309 "uuid": "85215d77-ec63-5266-b470-1eb93c194ef0", 00:08:04.309 "is_configured": true, 00:08:04.309 "data_offset": 2048, 00:08:04.309 "data_size": 63488 00:08:04.309 } 00:08:04.309 ] 00:08:04.309 }' 00:08:04.309 15:12:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:04.309 15:12:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:04.878 15:12:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:04.878 15:12:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:04.878 [2024-11-27 15:12:32.832053] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:08:05.816 15:12:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:08:05.816 15:12:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:05.816 15:12:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.816 15:12:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:05.816 15:12:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:05.816 15:12:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:08:05.816 15:12:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:08:05.816 15:12:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:08:05.816 15:12:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:08:05.816 15:12:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:05.816 15:12:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:05.816 15:12:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:05.816 15:12:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:05.816 15:12:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:05.816 15:12:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:05.816 15:12:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:05.816 15:12:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:05.816 15:12:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:05.816 15:12:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:05.816 15:12:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:05.816 15:12:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:05.816 15:12:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.816 15:12:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:05.816 15:12:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:05.816 "name": "raid_bdev1", 00:08:05.816 "uuid": "4b1f6f1a-5bf9-438d-82cc-45e8b27bba8e", 00:08:05.816 "strip_size_kb": 0, 00:08:05.816 "state": "online", 00:08:05.816 "raid_level": "raid1", 00:08:05.816 "superblock": true, 00:08:05.816 "num_base_bdevs": 2, 00:08:05.816 "num_base_bdevs_discovered": 2, 00:08:05.816 "num_base_bdevs_operational": 2, 00:08:05.816 "base_bdevs_list": [ 00:08:05.816 { 00:08:05.816 "name": "BaseBdev1", 00:08:05.816 "uuid": "8f27001d-e1a4-5694-b082-5cb006432068", 00:08:05.816 "is_configured": true, 00:08:05.816 "data_offset": 2048, 00:08:05.816 "data_size": 63488 00:08:05.816 }, 00:08:05.816 { 00:08:05.816 "name": "BaseBdev2", 00:08:05.816 "uuid": "85215d77-ec63-5266-b470-1eb93c194ef0", 00:08:05.816 "is_configured": true, 00:08:05.816 "data_offset": 2048, 00:08:05.816 "data_size": 63488 00:08:05.816 } 00:08:05.816 ] 00:08:05.816 }' 00:08:05.816 15:12:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:05.816 15:12:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.384 15:12:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:06.384 15:12:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.384 15:12:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.384 [2024-11-27 15:12:34.231729] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:06.384 [2024-11-27 15:12:34.231790] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:06.384 [2024-11-27 15:12:34.234376] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:06.384 [2024-11-27 15:12:34.234422] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:06.384 [2024-11-27 15:12:34.234504] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:06.384 [2024-11-27 15:12:34.234514] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:08:06.384 { 00:08:06.384 "results": [ 00:08:06.384 { 00:08:06.384 "job": "raid_bdev1", 00:08:06.384 "core_mask": "0x1", 00:08:06.384 "workload": "randrw", 00:08:06.384 "percentage": 50, 00:08:06.384 "status": "finished", 00:08:06.384 "queue_depth": 1, 00:08:06.384 "io_size": 131072, 00:08:06.384 "runtime": 1.400757, 00:08:06.384 "iops": 17682.581632645775, 00:08:06.384 "mibps": 2210.322704080722, 00:08:06.384 "io_failed": 0, 00:08:06.384 "io_timeout": 0, 00:08:06.384 "avg_latency_us": 53.960790331483864, 00:08:06.384 "min_latency_us": 22.358078602620086, 00:08:06.384 "max_latency_us": 1352.216593886463 00:08:06.384 } 00:08:06.384 ], 00:08:06.384 "core_count": 1 00:08:06.384 } 00:08:06.384 15:12:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.384 15:12:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 74939 00:08:06.384 15:12:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 74939 ']' 00:08:06.384 15:12:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 74939 00:08:06.384 15:12:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:08:06.384 15:12:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:06.384 15:12:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 74939 00:08:06.384 15:12:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:06.384 15:12:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:06.384 15:12:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 74939' 00:08:06.384 killing process with pid 74939 00:08:06.384 15:12:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 74939 00:08:06.384 [2024-11-27 15:12:34.286194] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:06.384 15:12:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 74939 00:08:06.384 [2024-11-27 15:12:34.302184] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:06.643 15:12:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.2QgjMJ8Fm2 00:08:06.643 15:12:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:06.643 15:12:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:06.643 15:12:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:08:06.643 15:12:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:08:06.643 15:12:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:06.643 15:12:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:08:06.643 15:12:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:08:06.643 00:08:06.643 real 0m3.272s 00:08:06.643 user 0m4.170s 00:08:06.643 sys 0m0.559s 00:08:06.643 15:12:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:06.643 15:12:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.643 ************************************ 00:08:06.643 END TEST raid_read_error_test 00:08:06.643 ************************************ 00:08:06.643 15:12:34 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 2 write 00:08:06.643 15:12:34 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:06.643 15:12:34 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:06.643 15:12:34 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:06.643 ************************************ 00:08:06.643 START TEST raid_write_error_test 00:08:06.644 ************************************ 00:08:06.644 15:12:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 2 write 00:08:06.644 15:12:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:08:06.644 15:12:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:08:06.644 15:12:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:08:06.644 15:12:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:06.644 15:12:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:06.644 15:12:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:06.644 15:12:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:06.644 15:12:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:06.644 15:12:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:06.644 15:12:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:06.644 15:12:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:06.644 15:12:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:08:06.644 15:12:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:06.644 15:12:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:06.644 15:12:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:06.644 15:12:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:06.644 15:12:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:06.644 15:12:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:06.644 15:12:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:08:06.644 15:12:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:08:06.644 15:12:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:06.644 15:12:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.py4b5xy2i3 00:08:06.644 15:12:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=75068 00:08:06.644 15:12:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 75068 00:08:06.644 15:12:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:06.644 15:12:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 75068 ']' 00:08:06.644 15:12:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:06.644 15:12:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:06.644 15:12:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:06.644 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:06.644 15:12:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:06.644 15:12:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.644 [2024-11-27 15:12:34.694466] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:08:06.644 [2024-11-27 15:12:34.695038] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid75068 ] 00:08:06.903 [2024-11-27 15:12:34.866605] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:06.903 [2024-11-27 15:12:34.897052] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:06.903 [2024-11-27 15:12:34.940371] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:06.903 [2024-11-27 15:12:34.940417] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:07.471 15:12:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:07.471 15:12:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:08:07.471 15:12:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:07.471 15:12:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:07.471 15:12:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:07.471 15:12:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.471 BaseBdev1_malloc 00:08:07.471 15:12:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:07.471 15:12:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:07.471 15:12:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:07.471 15:12:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.471 true 00:08:07.471 15:12:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:07.471 15:12:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:07.471 15:12:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:07.471 15:12:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.471 [2024-11-27 15:12:35.561244] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:07.471 [2024-11-27 15:12:35.561320] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:07.471 [2024-11-27 15:12:35.561346] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:08:07.471 [2024-11-27 15:12:35.561361] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:07.471 [2024-11-27 15:12:35.563684] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:07.471 [2024-11-27 15:12:35.563733] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:07.471 BaseBdev1 00:08:07.471 15:12:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:07.471 15:12:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:07.471 15:12:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:07.471 15:12:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:07.471 15:12:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.731 BaseBdev2_malloc 00:08:07.731 15:12:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:07.731 15:12:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:07.731 15:12:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:07.731 15:12:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.731 true 00:08:07.731 15:12:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:07.731 15:12:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:07.731 15:12:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:07.731 15:12:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.731 [2024-11-27 15:12:35.602791] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:07.731 [2024-11-27 15:12:35.602860] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:07.731 [2024-11-27 15:12:35.602900] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:08:07.731 [2024-11-27 15:12:35.602909] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:07.731 [2024-11-27 15:12:35.605267] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:07.731 [2024-11-27 15:12:35.605369] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:07.731 BaseBdev2 00:08:07.731 15:12:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:07.731 15:12:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:08:07.731 15:12:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:07.731 15:12:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.731 [2024-11-27 15:12:35.614841] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:07.731 [2024-11-27 15:12:35.617050] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:07.731 [2024-11-27 15:12:35.617249] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:08:07.731 [2024-11-27 15:12:35.617264] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:07.731 [2024-11-27 15:12:35.617581] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:08:07.731 [2024-11-27 15:12:35.617747] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:08:07.731 [2024-11-27 15:12:35.617765] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:08:07.731 [2024-11-27 15:12:35.617943] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:07.731 15:12:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:07.731 15:12:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:08:07.731 15:12:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:07.731 15:12:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:07.731 15:12:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:07.731 15:12:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:07.731 15:12:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:07.731 15:12:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:07.731 15:12:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:07.731 15:12:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:07.731 15:12:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:07.731 15:12:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:07.731 15:12:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:07.731 15:12:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:07.731 15:12:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.731 15:12:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:07.731 15:12:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:07.731 "name": "raid_bdev1", 00:08:07.731 "uuid": "96423769-8b80-4d3e-b715-1ceb58363de4", 00:08:07.731 "strip_size_kb": 0, 00:08:07.731 "state": "online", 00:08:07.731 "raid_level": "raid1", 00:08:07.731 "superblock": true, 00:08:07.731 "num_base_bdevs": 2, 00:08:07.731 "num_base_bdevs_discovered": 2, 00:08:07.731 "num_base_bdevs_operational": 2, 00:08:07.731 "base_bdevs_list": [ 00:08:07.731 { 00:08:07.731 "name": "BaseBdev1", 00:08:07.731 "uuid": "6b78d5a5-16ef-5f87-b484-b0864584d7cd", 00:08:07.731 "is_configured": true, 00:08:07.731 "data_offset": 2048, 00:08:07.731 "data_size": 63488 00:08:07.731 }, 00:08:07.731 { 00:08:07.731 "name": "BaseBdev2", 00:08:07.731 "uuid": "91f9bc7c-af85-5322-9018-bb6b86c01454", 00:08:07.731 "is_configured": true, 00:08:07.731 "data_offset": 2048, 00:08:07.731 "data_size": 63488 00:08:07.731 } 00:08:07.731 ] 00:08:07.731 }' 00:08:07.731 15:12:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:07.731 15:12:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.990 15:12:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:07.990 15:12:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:08.248 [2024-11-27 15:12:36.106347] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:08:09.186 15:12:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:08:09.186 15:12:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:09.186 15:12:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.186 [2024-11-27 15:12:37.023718] bdev_raid.c:2276:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:08:09.186 [2024-11-27 15:12:37.023930] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:09.186 [2024-11-27 15:12:37.024187] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000006220 00:08:09.186 15:12:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:09.186 15:12:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:09.186 15:12:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:08:09.186 15:12:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:08:09.186 15:12:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=1 00:08:09.186 15:12:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:08:09.186 15:12:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:09.186 15:12:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:09.186 15:12:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:09.186 15:12:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:09.186 15:12:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:08:09.186 15:12:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:09.186 15:12:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:09.186 15:12:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:09.186 15:12:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:09.186 15:12:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:09.186 15:12:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:09.186 15:12:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:09.186 15:12:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.186 15:12:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:09.186 15:12:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:09.186 "name": "raid_bdev1", 00:08:09.186 "uuid": "96423769-8b80-4d3e-b715-1ceb58363de4", 00:08:09.186 "strip_size_kb": 0, 00:08:09.186 "state": "online", 00:08:09.186 "raid_level": "raid1", 00:08:09.186 "superblock": true, 00:08:09.186 "num_base_bdevs": 2, 00:08:09.186 "num_base_bdevs_discovered": 1, 00:08:09.186 "num_base_bdevs_operational": 1, 00:08:09.186 "base_bdevs_list": [ 00:08:09.186 { 00:08:09.186 "name": null, 00:08:09.186 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:09.186 "is_configured": false, 00:08:09.186 "data_offset": 0, 00:08:09.186 "data_size": 63488 00:08:09.186 }, 00:08:09.186 { 00:08:09.186 "name": "BaseBdev2", 00:08:09.186 "uuid": "91f9bc7c-af85-5322-9018-bb6b86c01454", 00:08:09.186 "is_configured": true, 00:08:09.186 "data_offset": 2048, 00:08:09.186 "data_size": 63488 00:08:09.186 } 00:08:09.186 ] 00:08:09.186 }' 00:08:09.186 15:12:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:09.186 15:12:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.445 15:12:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:09.445 15:12:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:09.445 15:12:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.445 [2024-11-27 15:12:37.525348] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:09.445 [2024-11-27 15:12:37.525455] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:09.445 [2024-11-27 15:12:37.528054] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:09.445 [2024-11-27 15:12:37.528159] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:09.445 [2024-11-27 15:12:37.528234] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:09.445 [2024-11-27 15:12:37.528289] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:08:09.445 { 00:08:09.445 "results": [ 00:08:09.445 { 00:08:09.445 "job": "raid_bdev1", 00:08:09.445 "core_mask": "0x1", 00:08:09.445 "workload": "randrw", 00:08:09.445 "percentage": 50, 00:08:09.445 "status": "finished", 00:08:09.445 "queue_depth": 1, 00:08:09.445 "io_size": 131072, 00:08:09.445 "runtime": 1.419886, 00:08:09.445 "iops": 20743.214596101378, 00:08:09.445 "mibps": 2592.901824512672, 00:08:09.445 "io_failed": 0, 00:08:09.445 "io_timeout": 0, 00:08:09.445 "avg_latency_us": 45.6276324488264, 00:08:09.445 "min_latency_us": 21.910917030567685, 00:08:09.445 "max_latency_us": 1459.5353711790392 00:08:09.445 } 00:08:09.445 ], 00:08:09.445 "core_count": 1 00:08:09.445 } 00:08:09.445 15:12:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:09.445 15:12:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 75068 00:08:09.445 15:12:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 75068 ']' 00:08:09.445 15:12:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 75068 00:08:09.445 15:12:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:08:09.445 15:12:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:09.445 15:12:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 75068 00:08:09.704 killing process with pid 75068 00:08:09.704 15:12:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:09.704 15:12:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:09.704 15:12:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 75068' 00:08:09.704 15:12:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 75068 00:08:09.704 [2024-11-27 15:12:37.577780] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:09.704 15:12:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 75068 00:08:09.704 [2024-11-27 15:12:37.593750] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:09.704 15:12:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.py4b5xy2i3 00:08:09.704 15:12:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:09.704 15:12:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:09.962 15:12:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:08:09.962 15:12:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:08:09.962 15:12:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:09.962 15:12:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:08:09.962 15:12:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:08:09.962 00:08:09.962 real 0m3.223s 00:08:09.962 user 0m4.095s 00:08:09.962 sys 0m0.542s 00:08:09.962 15:12:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:09.962 ************************************ 00:08:09.962 END TEST raid_write_error_test 00:08:09.963 ************************************ 00:08:09.963 15:12:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.963 15:12:37 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:08:09.963 15:12:37 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:08:09.963 15:12:37 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 3 false 00:08:09.963 15:12:37 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:09.963 15:12:37 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:09.963 15:12:37 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:09.963 ************************************ 00:08:09.963 START TEST raid_state_function_test 00:08:09.963 ************************************ 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 3 false 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=75195 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 75195' 00:08:09.963 Process raid pid: 75195 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 75195 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 75195 ']' 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:09.963 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:09.963 15:12:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.963 [2024-11-27 15:12:37.986996] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:08:09.963 [2024-11-27 15:12:37.987202] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:10.221 [2024-11-27 15:12:38.159149] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:10.221 [2024-11-27 15:12:38.189431] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:10.221 [2024-11-27 15:12:38.233250] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:10.222 [2024-11-27 15:12:38.233420] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:10.789 15:12:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:10.789 15:12:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:08:10.789 15:12:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:10.789 15:12:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:10.789 15:12:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:10.789 [2024-11-27 15:12:38.825651] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:10.789 [2024-11-27 15:12:38.825765] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:10.789 [2024-11-27 15:12:38.825795] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:10.789 [2024-11-27 15:12:38.825818] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:10.789 [2024-11-27 15:12:38.825838] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:10.789 [2024-11-27 15:12:38.825863] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:10.789 15:12:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:10.789 15:12:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:10.789 15:12:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:10.789 15:12:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:10.789 15:12:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:10.789 15:12:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:10.789 15:12:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:10.789 15:12:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:10.789 15:12:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:10.789 15:12:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:10.789 15:12:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:10.789 15:12:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:10.789 15:12:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:10.789 15:12:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:10.789 15:12:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:10.789 15:12:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:10.789 15:12:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:10.789 "name": "Existed_Raid", 00:08:10.789 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:10.789 "strip_size_kb": 64, 00:08:10.789 "state": "configuring", 00:08:10.789 "raid_level": "raid0", 00:08:10.789 "superblock": false, 00:08:10.789 "num_base_bdevs": 3, 00:08:10.789 "num_base_bdevs_discovered": 0, 00:08:10.789 "num_base_bdevs_operational": 3, 00:08:10.789 "base_bdevs_list": [ 00:08:10.789 { 00:08:10.789 "name": "BaseBdev1", 00:08:10.789 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:10.789 "is_configured": false, 00:08:10.789 "data_offset": 0, 00:08:10.789 "data_size": 0 00:08:10.789 }, 00:08:10.789 { 00:08:10.789 "name": "BaseBdev2", 00:08:10.789 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:10.789 "is_configured": false, 00:08:10.789 "data_offset": 0, 00:08:10.789 "data_size": 0 00:08:10.789 }, 00:08:10.789 { 00:08:10.789 "name": "BaseBdev3", 00:08:10.789 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:10.789 "is_configured": false, 00:08:10.789 "data_offset": 0, 00:08:10.789 "data_size": 0 00:08:10.789 } 00:08:10.789 ] 00:08:10.789 }' 00:08:10.789 15:12:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:10.789 15:12:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.356 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:11.356 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:11.356 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.356 [2024-11-27 15:12:39.296778] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:11.356 [2024-11-27 15:12:39.296834] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:08:11.356 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:11.356 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:11.356 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:11.356 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.356 [2024-11-27 15:12:39.308780] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:11.356 [2024-11-27 15:12:39.308891] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:11.357 [2024-11-27 15:12:39.308936] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:11.357 [2024-11-27 15:12:39.308961] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:11.357 [2024-11-27 15:12:39.308992] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:11.357 [2024-11-27 15:12:39.309017] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.357 BaseBdev1 00:08:11.357 [2024-11-27 15:12:39.330070] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.357 [ 00:08:11.357 { 00:08:11.357 "name": "BaseBdev1", 00:08:11.357 "aliases": [ 00:08:11.357 "16459df1-1b8c-4cdd-a987-b3ef5d759221" 00:08:11.357 ], 00:08:11.357 "product_name": "Malloc disk", 00:08:11.357 "block_size": 512, 00:08:11.357 "num_blocks": 65536, 00:08:11.357 "uuid": "16459df1-1b8c-4cdd-a987-b3ef5d759221", 00:08:11.357 "assigned_rate_limits": { 00:08:11.357 "rw_ios_per_sec": 0, 00:08:11.357 "rw_mbytes_per_sec": 0, 00:08:11.357 "r_mbytes_per_sec": 0, 00:08:11.357 "w_mbytes_per_sec": 0 00:08:11.357 }, 00:08:11.357 "claimed": true, 00:08:11.357 "claim_type": "exclusive_write", 00:08:11.357 "zoned": false, 00:08:11.357 "supported_io_types": { 00:08:11.357 "read": true, 00:08:11.357 "write": true, 00:08:11.357 "unmap": true, 00:08:11.357 "flush": true, 00:08:11.357 "reset": true, 00:08:11.357 "nvme_admin": false, 00:08:11.357 "nvme_io": false, 00:08:11.357 "nvme_io_md": false, 00:08:11.357 "write_zeroes": true, 00:08:11.357 "zcopy": true, 00:08:11.357 "get_zone_info": false, 00:08:11.357 "zone_management": false, 00:08:11.357 "zone_append": false, 00:08:11.357 "compare": false, 00:08:11.357 "compare_and_write": false, 00:08:11.357 "abort": true, 00:08:11.357 "seek_hole": false, 00:08:11.357 "seek_data": false, 00:08:11.357 "copy": true, 00:08:11.357 "nvme_iov_md": false 00:08:11.357 }, 00:08:11.357 "memory_domains": [ 00:08:11.357 { 00:08:11.357 "dma_device_id": "system", 00:08:11.357 "dma_device_type": 1 00:08:11.357 }, 00:08:11.357 { 00:08:11.357 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:11.357 "dma_device_type": 2 00:08:11.357 } 00:08:11.357 ], 00:08:11.357 "driver_specific": {} 00:08:11.357 } 00:08:11.357 ] 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:11.357 "name": "Existed_Raid", 00:08:11.357 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:11.357 "strip_size_kb": 64, 00:08:11.357 "state": "configuring", 00:08:11.357 "raid_level": "raid0", 00:08:11.357 "superblock": false, 00:08:11.357 "num_base_bdevs": 3, 00:08:11.357 "num_base_bdevs_discovered": 1, 00:08:11.357 "num_base_bdevs_operational": 3, 00:08:11.357 "base_bdevs_list": [ 00:08:11.357 { 00:08:11.357 "name": "BaseBdev1", 00:08:11.357 "uuid": "16459df1-1b8c-4cdd-a987-b3ef5d759221", 00:08:11.357 "is_configured": true, 00:08:11.357 "data_offset": 0, 00:08:11.357 "data_size": 65536 00:08:11.357 }, 00:08:11.357 { 00:08:11.357 "name": "BaseBdev2", 00:08:11.357 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:11.357 "is_configured": false, 00:08:11.357 "data_offset": 0, 00:08:11.357 "data_size": 0 00:08:11.357 }, 00:08:11.357 { 00:08:11.357 "name": "BaseBdev3", 00:08:11.357 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:11.357 "is_configured": false, 00:08:11.357 "data_offset": 0, 00:08:11.357 "data_size": 0 00:08:11.357 } 00:08:11.357 ] 00:08:11.357 }' 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:11.357 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.926 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:11.926 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:11.926 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.926 [2024-11-27 15:12:39.789361] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:11.926 [2024-11-27 15:12:39.789503] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:08:11.926 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:11.926 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:11.926 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:11.926 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.926 [2024-11-27 15:12:39.801396] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:11.926 [2024-11-27 15:12:39.803336] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:11.926 [2024-11-27 15:12:39.803423] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:11.927 [2024-11-27 15:12:39.803452] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:11.927 [2024-11-27 15:12:39.803492] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:11.927 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:11.927 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:11.927 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:11.927 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:11.927 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:11.927 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:11.927 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:11.927 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:11.927 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:11.927 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:11.927 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:11.927 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:11.927 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:11.927 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:11.927 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:11.927 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.927 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:11.927 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:11.927 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:11.927 "name": "Existed_Raid", 00:08:11.927 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:11.927 "strip_size_kb": 64, 00:08:11.927 "state": "configuring", 00:08:11.927 "raid_level": "raid0", 00:08:11.927 "superblock": false, 00:08:11.927 "num_base_bdevs": 3, 00:08:11.927 "num_base_bdevs_discovered": 1, 00:08:11.927 "num_base_bdevs_operational": 3, 00:08:11.927 "base_bdevs_list": [ 00:08:11.927 { 00:08:11.927 "name": "BaseBdev1", 00:08:11.927 "uuid": "16459df1-1b8c-4cdd-a987-b3ef5d759221", 00:08:11.927 "is_configured": true, 00:08:11.927 "data_offset": 0, 00:08:11.927 "data_size": 65536 00:08:11.927 }, 00:08:11.927 { 00:08:11.927 "name": "BaseBdev2", 00:08:11.927 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:11.927 "is_configured": false, 00:08:11.927 "data_offset": 0, 00:08:11.927 "data_size": 0 00:08:11.927 }, 00:08:11.927 { 00:08:11.927 "name": "BaseBdev3", 00:08:11.927 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:11.927 "is_configured": false, 00:08:11.927 "data_offset": 0, 00:08:11.927 "data_size": 0 00:08:11.927 } 00:08:11.927 ] 00:08:11.927 }' 00:08:11.927 15:12:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:11.927 15:12:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.186 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:12.186 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:12.186 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.186 [2024-11-27 15:12:40.283855] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:12.186 BaseBdev2 00:08:12.186 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:12.186 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:12.186 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:12.186 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:12.186 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:12.186 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:12.186 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:12.186 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:12.186 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:12.186 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.446 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:12.446 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:12.446 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:12.446 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.446 [ 00:08:12.446 { 00:08:12.446 "name": "BaseBdev2", 00:08:12.446 "aliases": [ 00:08:12.446 "63222fc6-0b7f-4450-9844-5ee29c71f370" 00:08:12.446 ], 00:08:12.446 "product_name": "Malloc disk", 00:08:12.446 "block_size": 512, 00:08:12.446 "num_blocks": 65536, 00:08:12.446 "uuid": "63222fc6-0b7f-4450-9844-5ee29c71f370", 00:08:12.446 "assigned_rate_limits": { 00:08:12.446 "rw_ios_per_sec": 0, 00:08:12.446 "rw_mbytes_per_sec": 0, 00:08:12.446 "r_mbytes_per_sec": 0, 00:08:12.446 "w_mbytes_per_sec": 0 00:08:12.446 }, 00:08:12.446 "claimed": true, 00:08:12.446 "claim_type": "exclusive_write", 00:08:12.446 "zoned": false, 00:08:12.446 "supported_io_types": { 00:08:12.446 "read": true, 00:08:12.446 "write": true, 00:08:12.446 "unmap": true, 00:08:12.446 "flush": true, 00:08:12.446 "reset": true, 00:08:12.446 "nvme_admin": false, 00:08:12.446 "nvme_io": false, 00:08:12.446 "nvme_io_md": false, 00:08:12.446 "write_zeroes": true, 00:08:12.446 "zcopy": true, 00:08:12.446 "get_zone_info": false, 00:08:12.446 "zone_management": false, 00:08:12.446 "zone_append": false, 00:08:12.446 "compare": false, 00:08:12.446 "compare_and_write": false, 00:08:12.446 "abort": true, 00:08:12.446 "seek_hole": false, 00:08:12.446 "seek_data": false, 00:08:12.446 "copy": true, 00:08:12.446 "nvme_iov_md": false 00:08:12.446 }, 00:08:12.446 "memory_domains": [ 00:08:12.446 { 00:08:12.446 "dma_device_id": "system", 00:08:12.446 "dma_device_type": 1 00:08:12.446 }, 00:08:12.446 { 00:08:12.446 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:12.446 "dma_device_type": 2 00:08:12.446 } 00:08:12.446 ], 00:08:12.446 "driver_specific": {} 00:08:12.446 } 00:08:12.446 ] 00:08:12.446 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:12.446 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:12.446 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:12.446 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:12.446 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:12.446 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:12.446 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:12.446 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:12.446 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:12.446 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:12.446 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:12.446 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:12.446 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:12.446 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:12.446 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:12.446 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:12.446 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:12.446 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.446 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:12.446 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:12.446 "name": "Existed_Raid", 00:08:12.446 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:12.446 "strip_size_kb": 64, 00:08:12.446 "state": "configuring", 00:08:12.446 "raid_level": "raid0", 00:08:12.446 "superblock": false, 00:08:12.446 "num_base_bdevs": 3, 00:08:12.446 "num_base_bdevs_discovered": 2, 00:08:12.446 "num_base_bdevs_operational": 3, 00:08:12.446 "base_bdevs_list": [ 00:08:12.446 { 00:08:12.446 "name": "BaseBdev1", 00:08:12.446 "uuid": "16459df1-1b8c-4cdd-a987-b3ef5d759221", 00:08:12.446 "is_configured": true, 00:08:12.446 "data_offset": 0, 00:08:12.446 "data_size": 65536 00:08:12.446 }, 00:08:12.446 { 00:08:12.446 "name": "BaseBdev2", 00:08:12.446 "uuid": "63222fc6-0b7f-4450-9844-5ee29c71f370", 00:08:12.446 "is_configured": true, 00:08:12.446 "data_offset": 0, 00:08:12.446 "data_size": 65536 00:08:12.446 }, 00:08:12.446 { 00:08:12.446 "name": "BaseBdev3", 00:08:12.446 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:12.446 "is_configured": false, 00:08:12.446 "data_offset": 0, 00:08:12.446 "data_size": 0 00:08:12.446 } 00:08:12.446 ] 00:08:12.446 }' 00:08:12.446 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:12.446 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.706 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:12.706 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:12.706 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.706 [2024-11-27 15:12:40.787712] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:12.706 [2024-11-27 15:12:40.787894] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:08:12.706 [2024-11-27 15:12:40.787951] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:08:12.706 [2024-11-27 15:12:40.788347] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:08:12.706 [2024-11-27 15:12:40.788596] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:08:12.706 [2024-11-27 15:12:40.788648] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:08:12.706 [2024-11-27 15:12:40.788962] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:12.706 BaseBdev3 00:08:12.706 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:12.706 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:08:12.706 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:08:12.706 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:12.706 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:12.706 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:12.706 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:12.706 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:12.706 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:12.706 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.706 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:12.706 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:12.706 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:12.706 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.965 [ 00:08:12.965 { 00:08:12.965 "name": "BaseBdev3", 00:08:12.965 "aliases": [ 00:08:12.965 "84ee896a-2dba-4198-a010-30837d62b169" 00:08:12.965 ], 00:08:12.965 "product_name": "Malloc disk", 00:08:12.965 "block_size": 512, 00:08:12.965 "num_blocks": 65536, 00:08:12.965 "uuid": "84ee896a-2dba-4198-a010-30837d62b169", 00:08:12.965 "assigned_rate_limits": { 00:08:12.965 "rw_ios_per_sec": 0, 00:08:12.965 "rw_mbytes_per_sec": 0, 00:08:12.965 "r_mbytes_per_sec": 0, 00:08:12.965 "w_mbytes_per_sec": 0 00:08:12.965 }, 00:08:12.965 "claimed": true, 00:08:12.965 "claim_type": "exclusive_write", 00:08:12.965 "zoned": false, 00:08:12.965 "supported_io_types": { 00:08:12.965 "read": true, 00:08:12.965 "write": true, 00:08:12.965 "unmap": true, 00:08:12.965 "flush": true, 00:08:12.965 "reset": true, 00:08:12.965 "nvme_admin": false, 00:08:12.965 "nvme_io": false, 00:08:12.965 "nvme_io_md": false, 00:08:12.965 "write_zeroes": true, 00:08:12.965 "zcopy": true, 00:08:12.965 "get_zone_info": false, 00:08:12.965 "zone_management": false, 00:08:12.965 "zone_append": false, 00:08:12.965 "compare": false, 00:08:12.965 "compare_and_write": false, 00:08:12.965 "abort": true, 00:08:12.965 "seek_hole": false, 00:08:12.965 "seek_data": false, 00:08:12.965 "copy": true, 00:08:12.965 "nvme_iov_md": false 00:08:12.965 }, 00:08:12.965 "memory_domains": [ 00:08:12.965 { 00:08:12.965 "dma_device_id": "system", 00:08:12.965 "dma_device_type": 1 00:08:12.965 }, 00:08:12.965 { 00:08:12.965 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:12.965 "dma_device_type": 2 00:08:12.965 } 00:08:12.965 ], 00:08:12.965 "driver_specific": {} 00:08:12.965 } 00:08:12.965 ] 00:08:12.965 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:12.965 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:12.965 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:12.965 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:12.965 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:08:12.965 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:12.965 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:12.965 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:12.965 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:12.965 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:12.965 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:12.965 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:12.965 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:12.965 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:12.965 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:12.965 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:12.965 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:12.965 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.965 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:12.965 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:12.965 "name": "Existed_Raid", 00:08:12.965 "uuid": "9aedb762-f650-4802-bcec-1d018d18d4a5", 00:08:12.965 "strip_size_kb": 64, 00:08:12.965 "state": "online", 00:08:12.965 "raid_level": "raid0", 00:08:12.965 "superblock": false, 00:08:12.965 "num_base_bdevs": 3, 00:08:12.965 "num_base_bdevs_discovered": 3, 00:08:12.965 "num_base_bdevs_operational": 3, 00:08:12.965 "base_bdevs_list": [ 00:08:12.965 { 00:08:12.965 "name": "BaseBdev1", 00:08:12.965 "uuid": "16459df1-1b8c-4cdd-a987-b3ef5d759221", 00:08:12.965 "is_configured": true, 00:08:12.965 "data_offset": 0, 00:08:12.965 "data_size": 65536 00:08:12.965 }, 00:08:12.965 { 00:08:12.965 "name": "BaseBdev2", 00:08:12.965 "uuid": "63222fc6-0b7f-4450-9844-5ee29c71f370", 00:08:12.965 "is_configured": true, 00:08:12.965 "data_offset": 0, 00:08:12.965 "data_size": 65536 00:08:12.965 }, 00:08:12.965 { 00:08:12.965 "name": "BaseBdev3", 00:08:12.965 "uuid": "84ee896a-2dba-4198-a010-30837d62b169", 00:08:12.965 "is_configured": true, 00:08:12.965 "data_offset": 0, 00:08:12.965 "data_size": 65536 00:08:12.965 } 00:08:12.965 ] 00:08:12.965 }' 00:08:12.965 15:12:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:12.965 15:12:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.225 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:13.225 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:13.225 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:13.225 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:13.225 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:13.225 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:13.225 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:13.225 15:12:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:13.225 15:12:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.225 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:13.225 [2024-11-27 15:12:41.295214] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:13.225 15:12:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:13.485 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:13.485 "name": "Existed_Raid", 00:08:13.485 "aliases": [ 00:08:13.485 "9aedb762-f650-4802-bcec-1d018d18d4a5" 00:08:13.485 ], 00:08:13.485 "product_name": "Raid Volume", 00:08:13.485 "block_size": 512, 00:08:13.485 "num_blocks": 196608, 00:08:13.485 "uuid": "9aedb762-f650-4802-bcec-1d018d18d4a5", 00:08:13.485 "assigned_rate_limits": { 00:08:13.485 "rw_ios_per_sec": 0, 00:08:13.485 "rw_mbytes_per_sec": 0, 00:08:13.485 "r_mbytes_per_sec": 0, 00:08:13.485 "w_mbytes_per_sec": 0 00:08:13.485 }, 00:08:13.485 "claimed": false, 00:08:13.485 "zoned": false, 00:08:13.485 "supported_io_types": { 00:08:13.485 "read": true, 00:08:13.485 "write": true, 00:08:13.485 "unmap": true, 00:08:13.485 "flush": true, 00:08:13.485 "reset": true, 00:08:13.485 "nvme_admin": false, 00:08:13.485 "nvme_io": false, 00:08:13.485 "nvme_io_md": false, 00:08:13.485 "write_zeroes": true, 00:08:13.485 "zcopy": false, 00:08:13.485 "get_zone_info": false, 00:08:13.485 "zone_management": false, 00:08:13.485 "zone_append": false, 00:08:13.485 "compare": false, 00:08:13.485 "compare_and_write": false, 00:08:13.485 "abort": false, 00:08:13.485 "seek_hole": false, 00:08:13.485 "seek_data": false, 00:08:13.485 "copy": false, 00:08:13.485 "nvme_iov_md": false 00:08:13.485 }, 00:08:13.485 "memory_domains": [ 00:08:13.485 { 00:08:13.485 "dma_device_id": "system", 00:08:13.485 "dma_device_type": 1 00:08:13.485 }, 00:08:13.485 { 00:08:13.485 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:13.485 "dma_device_type": 2 00:08:13.485 }, 00:08:13.485 { 00:08:13.485 "dma_device_id": "system", 00:08:13.485 "dma_device_type": 1 00:08:13.485 }, 00:08:13.485 { 00:08:13.485 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:13.485 "dma_device_type": 2 00:08:13.485 }, 00:08:13.485 { 00:08:13.485 "dma_device_id": "system", 00:08:13.485 "dma_device_type": 1 00:08:13.485 }, 00:08:13.485 { 00:08:13.485 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:13.485 "dma_device_type": 2 00:08:13.485 } 00:08:13.485 ], 00:08:13.485 "driver_specific": { 00:08:13.485 "raid": { 00:08:13.485 "uuid": "9aedb762-f650-4802-bcec-1d018d18d4a5", 00:08:13.485 "strip_size_kb": 64, 00:08:13.485 "state": "online", 00:08:13.485 "raid_level": "raid0", 00:08:13.485 "superblock": false, 00:08:13.485 "num_base_bdevs": 3, 00:08:13.485 "num_base_bdevs_discovered": 3, 00:08:13.485 "num_base_bdevs_operational": 3, 00:08:13.485 "base_bdevs_list": [ 00:08:13.485 { 00:08:13.485 "name": "BaseBdev1", 00:08:13.485 "uuid": "16459df1-1b8c-4cdd-a987-b3ef5d759221", 00:08:13.485 "is_configured": true, 00:08:13.485 "data_offset": 0, 00:08:13.485 "data_size": 65536 00:08:13.485 }, 00:08:13.485 { 00:08:13.485 "name": "BaseBdev2", 00:08:13.485 "uuid": "63222fc6-0b7f-4450-9844-5ee29c71f370", 00:08:13.485 "is_configured": true, 00:08:13.485 "data_offset": 0, 00:08:13.485 "data_size": 65536 00:08:13.485 }, 00:08:13.485 { 00:08:13.485 "name": "BaseBdev3", 00:08:13.485 "uuid": "84ee896a-2dba-4198-a010-30837d62b169", 00:08:13.485 "is_configured": true, 00:08:13.485 "data_offset": 0, 00:08:13.485 "data_size": 65536 00:08:13.485 } 00:08:13.485 ] 00:08:13.485 } 00:08:13.485 } 00:08:13.485 }' 00:08:13.485 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:13.485 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:13.485 BaseBdev2 00:08:13.485 BaseBdev3' 00:08:13.485 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:13.485 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:13.485 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:13.485 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:13.485 15:12:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:13.485 15:12:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.485 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:13.485 15:12:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:13.485 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:13.485 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:13.485 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:13.485 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:13.485 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:13.485 15:12:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:13.485 15:12:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.485 15:12:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:13.485 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:13.485 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:13.485 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:13.485 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:13.485 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:13.485 15:12:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:13.485 15:12:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.485 15:12:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:13.485 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:13.485 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:13.485 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:13.485 15:12:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:13.485 15:12:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.745 [2024-11-27 15:12:41.590493] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:13.745 [2024-11-27 15:12:41.590584] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:13.745 [2024-11-27 15:12:41.590669] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:13.745 15:12:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:13.745 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:13.745 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:08:13.745 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:13.745 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:13.745 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:08:13.745 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 2 00:08:13.745 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:13.745 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:08:13.745 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:13.745 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:13.745 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:13.745 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:13.745 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:13.745 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:13.745 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:13.745 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:13.745 15:12:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:13.745 15:12:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.745 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:13.745 15:12:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:13.745 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:13.745 "name": "Existed_Raid", 00:08:13.745 "uuid": "9aedb762-f650-4802-bcec-1d018d18d4a5", 00:08:13.745 "strip_size_kb": 64, 00:08:13.745 "state": "offline", 00:08:13.745 "raid_level": "raid0", 00:08:13.745 "superblock": false, 00:08:13.745 "num_base_bdevs": 3, 00:08:13.745 "num_base_bdevs_discovered": 2, 00:08:13.745 "num_base_bdevs_operational": 2, 00:08:13.745 "base_bdevs_list": [ 00:08:13.745 { 00:08:13.745 "name": null, 00:08:13.745 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:13.745 "is_configured": false, 00:08:13.745 "data_offset": 0, 00:08:13.745 "data_size": 65536 00:08:13.745 }, 00:08:13.745 { 00:08:13.745 "name": "BaseBdev2", 00:08:13.745 "uuid": "63222fc6-0b7f-4450-9844-5ee29c71f370", 00:08:13.745 "is_configured": true, 00:08:13.745 "data_offset": 0, 00:08:13.745 "data_size": 65536 00:08:13.745 }, 00:08:13.745 { 00:08:13.745 "name": "BaseBdev3", 00:08:13.745 "uuid": "84ee896a-2dba-4198-a010-30837d62b169", 00:08:13.745 "is_configured": true, 00:08:13.745 "data_offset": 0, 00:08:13.745 "data_size": 65536 00:08:13.745 } 00:08:13.745 ] 00:08:13.745 }' 00:08:13.745 15:12:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:13.745 15:12:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.004 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:14.004 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:14.004 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:14.004 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:14.004 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.004 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:14.004 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:14.004 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:14.004 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:14.004 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:14.004 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:14.004 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.004 [2024-11-27 15:12:42.093565] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:14.004 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:14.004 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:14.004 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.264 [2024-11-27 15:12:42.148787] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:14.264 [2024-11-27 15:12:42.148879] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.264 BaseBdev2 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.264 [ 00:08:14.264 { 00:08:14.264 "name": "BaseBdev2", 00:08:14.264 "aliases": [ 00:08:14.264 "208756dc-8205-49ce-b225-6205e0673803" 00:08:14.264 ], 00:08:14.264 "product_name": "Malloc disk", 00:08:14.264 "block_size": 512, 00:08:14.264 "num_blocks": 65536, 00:08:14.264 "uuid": "208756dc-8205-49ce-b225-6205e0673803", 00:08:14.264 "assigned_rate_limits": { 00:08:14.264 "rw_ios_per_sec": 0, 00:08:14.264 "rw_mbytes_per_sec": 0, 00:08:14.264 "r_mbytes_per_sec": 0, 00:08:14.264 "w_mbytes_per_sec": 0 00:08:14.264 }, 00:08:14.264 "claimed": false, 00:08:14.264 "zoned": false, 00:08:14.264 "supported_io_types": { 00:08:14.264 "read": true, 00:08:14.264 "write": true, 00:08:14.264 "unmap": true, 00:08:14.264 "flush": true, 00:08:14.264 "reset": true, 00:08:14.264 "nvme_admin": false, 00:08:14.264 "nvme_io": false, 00:08:14.264 "nvme_io_md": false, 00:08:14.264 "write_zeroes": true, 00:08:14.264 "zcopy": true, 00:08:14.264 "get_zone_info": false, 00:08:14.264 "zone_management": false, 00:08:14.264 "zone_append": false, 00:08:14.264 "compare": false, 00:08:14.264 "compare_and_write": false, 00:08:14.264 "abort": true, 00:08:14.264 "seek_hole": false, 00:08:14.264 "seek_data": false, 00:08:14.264 "copy": true, 00:08:14.264 "nvme_iov_md": false 00:08:14.264 }, 00:08:14.264 "memory_domains": [ 00:08:14.264 { 00:08:14.264 "dma_device_id": "system", 00:08:14.264 "dma_device_type": 1 00:08:14.264 }, 00:08:14.264 { 00:08:14.264 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:14.264 "dma_device_type": 2 00:08:14.264 } 00:08:14.264 ], 00:08:14.264 "driver_specific": {} 00:08:14.264 } 00:08:14.264 ] 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.264 BaseBdev3 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.264 [ 00:08:14.264 { 00:08:14.264 "name": "BaseBdev3", 00:08:14.264 "aliases": [ 00:08:14.264 "9d74845d-25c2-4fe6-bb22-364f8807162f" 00:08:14.264 ], 00:08:14.264 "product_name": "Malloc disk", 00:08:14.264 "block_size": 512, 00:08:14.264 "num_blocks": 65536, 00:08:14.264 "uuid": "9d74845d-25c2-4fe6-bb22-364f8807162f", 00:08:14.264 "assigned_rate_limits": { 00:08:14.264 "rw_ios_per_sec": 0, 00:08:14.264 "rw_mbytes_per_sec": 0, 00:08:14.264 "r_mbytes_per_sec": 0, 00:08:14.264 "w_mbytes_per_sec": 0 00:08:14.264 }, 00:08:14.264 "claimed": false, 00:08:14.264 "zoned": false, 00:08:14.264 "supported_io_types": { 00:08:14.264 "read": true, 00:08:14.264 "write": true, 00:08:14.264 "unmap": true, 00:08:14.264 "flush": true, 00:08:14.264 "reset": true, 00:08:14.264 "nvme_admin": false, 00:08:14.264 "nvme_io": false, 00:08:14.264 "nvme_io_md": false, 00:08:14.264 "write_zeroes": true, 00:08:14.264 "zcopy": true, 00:08:14.264 "get_zone_info": false, 00:08:14.264 "zone_management": false, 00:08:14.264 "zone_append": false, 00:08:14.264 "compare": false, 00:08:14.264 "compare_and_write": false, 00:08:14.264 "abort": true, 00:08:14.264 "seek_hole": false, 00:08:14.264 "seek_data": false, 00:08:14.264 "copy": true, 00:08:14.264 "nvme_iov_md": false 00:08:14.264 }, 00:08:14.264 "memory_domains": [ 00:08:14.264 { 00:08:14.264 "dma_device_id": "system", 00:08:14.264 "dma_device_type": 1 00:08:14.264 }, 00:08:14.264 { 00:08:14.264 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:14.264 "dma_device_type": 2 00:08:14.264 } 00:08:14.264 ], 00:08:14.264 "driver_specific": {} 00:08:14.264 } 00:08:14.264 ] 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.264 [2024-11-27 15:12:42.325772] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:14.264 [2024-11-27 15:12:42.325866] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:14.264 [2024-11-27 15:12:42.325918] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:14.264 [2024-11-27 15:12:42.327784] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.264 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:14.524 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:14.524 "name": "Existed_Raid", 00:08:14.524 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:14.524 "strip_size_kb": 64, 00:08:14.524 "state": "configuring", 00:08:14.524 "raid_level": "raid0", 00:08:14.524 "superblock": false, 00:08:14.524 "num_base_bdevs": 3, 00:08:14.524 "num_base_bdevs_discovered": 2, 00:08:14.524 "num_base_bdevs_operational": 3, 00:08:14.524 "base_bdevs_list": [ 00:08:14.524 { 00:08:14.524 "name": "BaseBdev1", 00:08:14.524 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:14.524 "is_configured": false, 00:08:14.524 "data_offset": 0, 00:08:14.524 "data_size": 0 00:08:14.524 }, 00:08:14.524 { 00:08:14.524 "name": "BaseBdev2", 00:08:14.524 "uuid": "208756dc-8205-49ce-b225-6205e0673803", 00:08:14.524 "is_configured": true, 00:08:14.524 "data_offset": 0, 00:08:14.524 "data_size": 65536 00:08:14.524 }, 00:08:14.524 { 00:08:14.524 "name": "BaseBdev3", 00:08:14.524 "uuid": "9d74845d-25c2-4fe6-bb22-364f8807162f", 00:08:14.524 "is_configured": true, 00:08:14.524 "data_offset": 0, 00:08:14.524 "data_size": 65536 00:08:14.524 } 00:08:14.524 ] 00:08:14.524 }' 00:08:14.524 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:14.524 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.829 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:08:14.829 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:14.829 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.830 [2024-11-27 15:12:42.773031] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:14.830 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:14.830 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:14.830 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:14.830 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:14.830 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:14.830 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:14.830 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:14.830 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:14.830 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:14.830 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:14.830 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:14.830 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:14.830 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:14.830 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:14.830 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.830 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:14.830 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:14.830 "name": "Existed_Raid", 00:08:14.830 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:14.830 "strip_size_kb": 64, 00:08:14.830 "state": "configuring", 00:08:14.830 "raid_level": "raid0", 00:08:14.830 "superblock": false, 00:08:14.830 "num_base_bdevs": 3, 00:08:14.830 "num_base_bdevs_discovered": 1, 00:08:14.830 "num_base_bdevs_operational": 3, 00:08:14.830 "base_bdevs_list": [ 00:08:14.830 { 00:08:14.830 "name": "BaseBdev1", 00:08:14.830 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:14.830 "is_configured": false, 00:08:14.830 "data_offset": 0, 00:08:14.830 "data_size": 0 00:08:14.830 }, 00:08:14.830 { 00:08:14.830 "name": null, 00:08:14.830 "uuid": "208756dc-8205-49ce-b225-6205e0673803", 00:08:14.830 "is_configured": false, 00:08:14.830 "data_offset": 0, 00:08:14.830 "data_size": 65536 00:08:14.830 }, 00:08:14.830 { 00:08:14.830 "name": "BaseBdev3", 00:08:14.830 "uuid": "9d74845d-25c2-4fe6-bb22-364f8807162f", 00:08:14.830 "is_configured": true, 00:08:14.830 "data_offset": 0, 00:08:14.830 "data_size": 65536 00:08:14.830 } 00:08:14.830 ] 00:08:14.830 }' 00:08:14.830 15:12:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:14.830 15:12:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.398 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:15.398 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:15.398 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.398 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:15.398 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:15.398 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:08:15.398 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:15.398 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:15.398 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.398 [2024-11-27 15:12:43.259331] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:15.398 BaseBdev1 00:08:15.398 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:15.398 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:08:15.398 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:15.398 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:15.398 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:15.398 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:15.398 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:15.398 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:15.398 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:15.398 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.398 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:15.398 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:15.398 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:15.398 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.398 [ 00:08:15.398 { 00:08:15.398 "name": "BaseBdev1", 00:08:15.398 "aliases": [ 00:08:15.398 "b5b8bd93-8893-47ba-8db7-1f8101924ec6" 00:08:15.398 ], 00:08:15.398 "product_name": "Malloc disk", 00:08:15.398 "block_size": 512, 00:08:15.398 "num_blocks": 65536, 00:08:15.398 "uuid": "b5b8bd93-8893-47ba-8db7-1f8101924ec6", 00:08:15.398 "assigned_rate_limits": { 00:08:15.398 "rw_ios_per_sec": 0, 00:08:15.398 "rw_mbytes_per_sec": 0, 00:08:15.398 "r_mbytes_per_sec": 0, 00:08:15.398 "w_mbytes_per_sec": 0 00:08:15.398 }, 00:08:15.398 "claimed": true, 00:08:15.398 "claim_type": "exclusive_write", 00:08:15.398 "zoned": false, 00:08:15.398 "supported_io_types": { 00:08:15.398 "read": true, 00:08:15.398 "write": true, 00:08:15.398 "unmap": true, 00:08:15.398 "flush": true, 00:08:15.398 "reset": true, 00:08:15.398 "nvme_admin": false, 00:08:15.398 "nvme_io": false, 00:08:15.398 "nvme_io_md": false, 00:08:15.398 "write_zeroes": true, 00:08:15.398 "zcopy": true, 00:08:15.398 "get_zone_info": false, 00:08:15.398 "zone_management": false, 00:08:15.398 "zone_append": false, 00:08:15.398 "compare": false, 00:08:15.398 "compare_and_write": false, 00:08:15.398 "abort": true, 00:08:15.398 "seek_hole": false, 00:08:15.398 "seek_data": false, 00:08:15.398 "copy": true, 00:08:15.398 "nvme_iov_md": false 00:08:15.398 }, 00:08:15.398 "memory_domains": [ 00:08:15.398 { 00:08:15.398 "dma_device_id": "system", 00:08:15.398 "dma_device_type": 1 00:08:15.399 }, 00:08:15.399 { 00:08:15.399 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:15.399 "dma_device_type": 2 00:08:15.399 } 00:08:15.399 ], 00:08:15.399 "driver_specific": {} 00:08:15.399 } 00:08:15.399 ] 00:08:15.399 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:15.399 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:15.399 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:15.399 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:15.399 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:15.399 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:15.399 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:15.399 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:15.399 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:15.399 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:15.399 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:15.399 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:15.399 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:15.399 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:15.399 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:15.399 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.399 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:15.399 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:15.399 "name": "Existed_Raid", 00:08:15.399 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:15.399 "strip_size_kb": 64, 00:08:15.399 "state": "configuring", 00:08:15.399 "raid_level": "raid0", 00:08:15.399 "superblock": false, 00:08:15.399 "num_base_bdevs": 3, 00:08:15.399 "num_base_bdevs_discovered": 2, 00:08:15.399 "num_base_bdevs_operational": 3, 00:08:15.399 "base_bdevs_list": [ 00:08:15.399 { 00:08:15.399 "name": "BaseBdev1", 00:08:15.399 "uuid": "b5b8bd93-8893-47ba-8db7-1f8101924ec6", 00:08:15.399 "is_configured": true, 00:08:15.399 "data_offset": 0, 00:08:15.399 "data_size": 65536 00:08:15.399 }, 00:08:15.399 { 00:08:15.399 "name": null, 00:08:15.399 "uuid": "208756dc-8205-49ce-b225-6205e0673803", 00:08:15.399 "is_configured": false, 00:08:15.399 "data_offset": 0, 00:08:15.399 "data_size": 65536 00:08:15.399 }, 00:08:15.399 { 00:08:15.399 "name": "BaseBdev3", 00:08:15.399 "uuid": "9d74845d-25c2-4fe6-bb22-364f8807162f", 00:08:15.399 "is_configured": true, 00:08:15.399 "data_offset": 0, 00:08:15.399 "data_size": 65536 00:08:15.399 } 00:08:15.399 ] 00:08:15.399 }' 00:08:15.399 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:15.399 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.668 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:15.668 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:15.668 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:15.668 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.668 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:15.945 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:08:15.945 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:08:15.945 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:15.945 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.945 [2024-11-27 15:12:43.790534] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:15.945 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:15.946 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:15.946 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:15.946 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:15.946 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:15.946 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:15.946 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:15.946 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:15.946 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:15.946 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:15.946 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:15.946 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:15.946 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:15.946 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:15.946 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.946 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:15.946 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:15.946 "name": "Existed_Raid", 00:08:15.946 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:15.946 "strip_size_kb": 64, 00:08:15.946 "state": "configuring", 00:08:15.946 "raid_level": "raid0", 00:08:15.946 "superblock": false, 00:08:15.946 "num_base_bdevs": 3, 00:08:15.946 "num_base_bdevs_discovered": 1, 00:08:15.946 "num_base_bdevs_operational": 3, 00:08:15.946 "base_bdevs_list": [ 00:08:15.946 { 00:08:15.946 "name": "BaseBdev1", 00:08:15.946 "uuid": "b5b8bd93-8893-47ba-8db7-1f8101924ec6", 00:08:15.946 "is_configured": true, 00:08:15.946 "data_offset": 0, 00:08:15.946 "data_size": 65536 00:08:15.946 }, 00:08:15.946 { 00:08:15.946 "name": null, 00:08:15.946 "uuid": "208756dc-8205-49ce-b225-6205e0673803", 00:08:15.946 "is_configured": false, 00:08:15.946 "data_offset": 0, 00:08:15.946 "data_size": 65536 00:08:15.946 }, 00:08:15.946 { 00:08:15.946 "name": null, 00:08:15.946 "uuid": "9d74845d-25c2-4fe6-bb22-364f8807162f", 00:08:15.946 "is_configured": false, 00:08:15.946 "data_offset": 0, 00:08:15.946 "data_size": 65536 00:08:15.946 } 00:08:15.946 ] 00:08:15.946 }' 00:08:15.946 15:12:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:15.946 15:12:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.215 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:16.215 15:12:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.215 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:16.215 15:12:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.215 15:12:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.215 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:08:16.215 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:08:16.215 15:12:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.215 15:12:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.215 [2024-11-27 15:12:44.309587] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:16.215 15:12:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.215 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:16.215 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:16.215 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:16.215 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:16.215 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:16.215 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:16.215 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:16.215 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:16.215 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:16.215 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:16.474 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:16.474 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:16.474 15:12:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.474 15:12:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.474 15:12:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.474 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:16.474 "name": "Existed_Raid", 00:08:16.474 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:16.474 "strip_size_kb": 64, 00:08:16.474 "state": "configuring", 00:08:16.474 "raid_level": "raid0", 00:08:16.474 "superblock": false, 00:08:16.474 "num_base_bdevs": 3, 00:08:16.474 "num_base_bdevs_discovered": 2, 00:08:16.474 "num_base_bdevs_operational": 3, 00:08:16.474 "base_bdevs_list": [ 00:08:16.474 { 00:08:16.474 "name": "BaseBdev1", 00:08:16.474 "uuid": "b5b8bd93-8893-47ba-8db7-1f8101924ec6", 00:08:16.474 "is_configured": true, 00:08:16.474 "data_offset": 0, 00:08:16.474 "data_size": 65536 00:08:16.474 }, 00:08:16.474 { 00:08:16.474 "name": null, 00:08:16.474 "uuid": "208756dc-8205-49ce-b225-6205e0673803", 00:08:16.474 "is_configured": false, 00:08:16.474 "data_offset": 0, 00:08:16.474 "data_size": 65536 00:08:16.474 }, 00:08:16.474 { 00:08:16.474 "name": "BaseBdev3", 00:08:16.474 "uuid": "9d74845d-25c2-4fe6-bb22-364f8807162f", 00:08:16.474 "is_configured": true, 00:08:16.474 "data_offset": 0, 00:08:16.474 "data_size": 65536 00:08:16.474 } 00:08:16.474 ] 00:08:16.474 }' 00:08:16.474 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:16.474 15:12:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.733 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:16.733 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:16.733 15:12:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.733 15:12:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.733 15:12:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.733 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:08:16.733 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:16.733 15:12:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.733 15:12:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.733 [2024-11-27 15:12:44.820812] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:16.733 15:12:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.733 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:16.733 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:16.733 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:16.733 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:16.733 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:16.733 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:16.733 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:16.733 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:16.733 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:16.733 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:16.991 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:16.991 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:16.991 15:12:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.991 15:12:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.992 15:12:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.992 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:16.992 "name": "Existed_Raid", 00:08:16.992 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:16.992 "strip_size_kb": 64, 00:08:16.992 "state": "configuring", 00:08:16.992 "raid_level": "raid0", 00:08:16.992 "superblock": false, 00:08:16.992 "num_base_bdevs": 3, 00:08:16.992 "num_base_bdevs_discovered": 1, 00:08:16.992 "num_base_bdevs_operational": 3, 00:08:16.992 "base_bdevs_list": [ 00:08:16.992 { 00:08:16.992 "name": null, 00:08:16.992 "uuid": "b5b8bd93-8893-47ba-8db7-1f8101924ec6", 00:08:16.992 "is_configured": false, 00:08:16.992 "data_offset": 0, 00:08:16.992 "data_size": 65536 00:08:16.992 }, 00:08:16.992 { 00:08:16.992 "name": null, 00:08:16.992 "uuid": "208756dc-8205-49ce-b225-6205e0673803", 00:08:16.992 "is_configured": false, 00:08:16.992 "data_offset": 0, 00:08:16.992 "data_size": 65536 00:08:16.992 }, 00:08:16.992 { 00:08:16.992 "name": "BaseBdev3", 00:08:16.992 "uuid": "9d74845d-25c2-4fe6-bb22-364f8807162f", 00:08:16.992 "is_configured": true, 00:08:16.992 "data_offset": 0, 00:08:16.992 "data_size": 65536 00:08:16.992 } 00:08:16.992 ] 00:08:16.992 }' 00:08:16.992 15:12:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:16.992 15:12:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.250 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:17.250 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:17.250 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.250 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.250 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.509 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:08:17.509 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:08:17.509 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.509 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.509 [2024-11-27 15:12:45.370573] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:17.509 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.509 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:17.509 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:17.509 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:17.509 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:17.509 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:17.509 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:17.509 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:17.509 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:17.509 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:17.509 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:17.509 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:17.509 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:17.509 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.509 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.509 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.509 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:17.509 "name": "Existed_Raid", 00:08:17.509 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:17.509 "strip_size_kb": 64, 00:08:17.509 "state": "configuring", 00:08:17.509 "raid_level": "raid0", 00:08:17.509 "superblock": false, 00:08:17.509 "num_base_bdevs": 3, 00:08:17.509 "num_base_bdevs_discovered": 2, 00:08:17.509 "num_base_bdevs_operational": 3, 00:08:17.509 "base_bdevs_list": [ 00:08:17.509 { 00:08:17.509 "name": null, 00:08:17.509 "uuid": "b5b8bd93-8893-47ba-8db7-1f8101924ec6", 00:08:17.509 "is_configured": false, 00:08:17.509 "data_offset": 0, 00:08:17.509 "data_size": 65536 00:08:17.509 }, 00:08:17.509 { 00:08:17.509 "name": "BaseBdev2", 00:08:17.509 "uuid": "208756dc-8205-49ce-b225-6205e0673803", 00:08:17.509 "is_configured": true, 00:08:17.509 "data_offset": 0, 00:08:17.509 "data_size": 65536 00:08:17.509 }, 00:08:17.509 { 00:08:17.509 "name": "BaseBdev3", 00:08:17.509 "uuid": "9d74845d-25c2-4fe6-bb22-364f8807162f", 00:08:17.509 "is_configured": true, 00:08:17.509 "data_offset": 0, 00:08:17.509 "data_size": 65536 00:08:17.509 } 00:08:17.509 ] 00:08:17.509 }' 00:08:17.509 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:17.509 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.768 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:17.768 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:17.768 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.768 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.768 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.768 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:08:17.768 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:17.768 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.768 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.768 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:08:17.768 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.768 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u b5b8bd93-8893-47ba-8db7-1f8101924ec6 00:08:17.768 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.768 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.028 [2024-11-27 15:12:45.884989] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:08:18.028 [2024-11-27 15:12:45.885104] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:08:18.028 [2024-11-27 15:12:45.885132] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:08:18.028 [2024-11-27 15:12:45.885421] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:08:18.028 [2024-11-27 15:12:45.885585] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:08:18.028 [2024-11-27 15:12:45.885626] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006d00 00:08:18.028 [2024-11-27 15:12:45.885841] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:18.028 NewBaseBdev 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.028 [ 00:08:18.028 { 00:08:18.028 "name": "NewBaseBdev", 00:08:18.028 "aliases": [ 00:08:18.028 "b5b8bd93-8893-47ba-8db7-1f8101924ec6" 00:08:18.028 ], 00:08:18.028 "product_name": "Malloc disk", 00:08:18.028 "block_size": 512, 00:08:18.028 "num_blocks": 65536, 00:08:18.028 "uuid": "b5b8bd93-8893-47ba-8db7-1f8101924ec6", 00:08:18.028 "assigned_rate_limits": { 00:08:18.028 "rw_ios_per_sec": 0, 00:08:18.028 "rw_mbytes_per_sec": 0, 00:08:18.028 "r_mbytes_per_sec": 0, 00:08:18.028 "w_mbytes_per_sec": 0 00:08:18.028 }, 00:08:18.028 "claimed": true, 00:08:18.028 "claim_type": "exclusive_write", 00:08:18.028 "zoned": false, 00:08:18.028 "supported_io_types": { 00:08:18.028 "read": true, 00:08:18.028 "write": true, 00:08:18.028 "unmap": true, 00:08:18.028 "flush": true, 00:08:18.028 "reset": true, 00:08:18.028 "nvme_admin": false, 00:08:18.028 "nvme_io": false, 00:08:18.028 "nvme_io_md": false, 00:08:18.028 "write_zeroes": true, 00:08:18.028 "zcopy": true, 00:08:18.028 "get_zone_info": false, 00:08:18.028 "zone_management": false, 00:08:18.028 "zone_append": false, 00:08:18.028 "compare": false, 00:08:18.028 "compare_and_write": false, 00:08:18.028 "abort": true, 00:08:18.028 "seek_hole": false, 00:08:18.028 "seek_data": false, 00:08:18.028 "copy": true, 00:08:18.028 "nvme_iov_md": false 00:08:18.028 }, 00:08:18.028 "memory_domains": [ 00:08:18.028 { 00:08:18.028 "dma_device_id": "system", 00:08:18.028 "dma_device_type": 1 00:08:18.028 }, 00:08:18.028 { 00:08:18.028 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:18.028 "dma_device_type": 2 00:08:18.028 } 00:08:18.028 ], 00:08:18.028 "driver_specific": {} 00:08:18.028 } 00:08:18.028 ] 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:18.028 "name": "Existed_Raid", 00:08:18.028 "uuid": "74717d81-5982-4287-9c73-f2af742ccddd", 00:08:18.028 "strip_size_kb": 64, 00:08:18.028 "state": "online", 00:08:18.028 "raid_level": "raid0", 00:08:18.028 "superblock": false, 00:08:18.028 "num_base_bdevs": 3, 00:08:18.028 "num_base_bdevs_discovered": 3, 00:08:18.028 "num_base_bdevs_operational": 3, 00:08:18.028 "base_bdevs_list": [ 00:08:18.028 { 00:08:18.028 "name": "NewBaseBdev", 00:08:18.028 "uuid": "b5b8bd93-8893-47ba-8db7-1f8101924ec6", 00:08:18.028 "is_configured": true, 00:08:18.028 "data_offset": 0, 00:08:18.028 "data_size": 65536 00:08:18.028 }, 00:08:18.028 { 00:08:18.028 "name": "BaseBdev2", 00:08:18.028 "uuid": "208756dc-8205-49ce-b225-6205e0673803", 00:08:18.028 "is_configured": true, 00:08:18.028 "data_offset": 0, 00:08:18.028 "data_size": 65536 00:08:18.028 }, 00:08:18.028 { 00:08:18.028 "name": "BaseBdev3", 00:08:18.028 "uuid": "9d74845d-25c2-4fe6-bb22-364f8807162f", 00:08:18.028 "is_configured": true, 00:08:18.028 "data_offset": 0, 00:08:18.028 "data_size": 65536 00:08:18.028 } 00:08:18.028 ] 00:08:18.028 }' 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:18.028 15:12:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.288 15:12:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:08:18.288 15:12:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:18.288 15:12:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:18.288 15:12:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:18.288 15:12:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:18.288 15:12:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:18.288 15:12:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:18.288 15:12:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:18.288 15:12:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.288 15:12:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.288 [2024-11-27 15:12:46.388493] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:18.547 15:12:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.547 15:12:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:18.547 "name": "Existed_Raid", 00:08:18.547 "aliases": [ 00:08:18.547 "74717d81-5982-4287-9c73-f2af742ccddd" 00:08:18.547 ], 00:08:18.547 "product_name": "Raid Volume", 00:08:18.547 "block_size": 512, 00:08:18.547 "num_blocks": 196608, 00:08:18.547 "uuid": "74717d81-5982-4287-9c73-f2af742ccddd", 00:08:18.547 "assigned_rate_limits": { 00:08:18.547 "rw_ios_per_sec": 0, 00:08:18.547 "rw_mbytes_per_sec": 0, 00:08:18.547 "r_mbytes_per_sec": 0, 00:08:18.547 "w_mbytes_per_sec": 0 00:08:18.547 }, 00:08:18.547 "claimed": false, 00:08:18.547 "zoned": false, 00:08:18.547 "supported_io_types": { 00:08:18.547 "read": true, 00:08:18.547 "write": true, 00:08:18.547 "unmap": true, 00:08:18.547 "flush": true, 00:08:18.547 "reset": true, 00:08:18.547 "nvme_admin": false, 00:08:18.547 "nvme_io": false, 00:08:18.547 "nvme_io_md": false, 00:08:18.547 "write_zeroes": true, 00:08:18.547 "zcopy": false, 00:08:18.547 "get_zone_info": false, 00:08:18.547 "zone_management": false, 00:08:18.547 "zone_append": false, 00:08:18.547 "compare": false, 00:08:18.547 "compare_and_write": false, 00:08:18.547 "abort": false, 00:08:18.547 "seek_hole": false, 00:08:18.547 "seek_data": false, 00:08:18.547 "copy": false, 00:08:18.547 "nvme_iov_md": false 00:08:18.547 }, 00:08:18.547 "memory_domains": [ 00:08:18.547 { 00:08:18.547 "dma_device_id": "system", 00:08:18.547 "dma_device_type": 1 00:08:18.547 }, 00:08:18.547 { 00:08:18.547 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:18.547 "dma_device_type": 2 00:08:18.547 }, 00:08:18.547 { 00:08:18.547 "dma_device_id": "system", 00:08:18.547 "dma_device_type": 1 00:08:18.547 }, 00:08:18.547 { 00:08:18.547 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:18.547 "dma_device_type": 2 00:08:18.547 }, 00:08:18.547 { 00:08:18.547 "dma_device_id": "system", 00:08:18.547 "dma_device_type": 1 00:08:18.547 }, 00:08:18.547 { 00:08:18.547 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:18.547 "dma_device_type": 2 00:08:18.547 } 00:08:18.547 ], 00:08:18.547 "driver_specific": { 00:08:18.547 "raid": { 00:08:18.547 "uuid": "74717d81-5982-4287-9c73-f2af742ccddd", 00:08:18.547 "strip_size_kb": 64, 00:08:18.547 "state": "online", 00:08:18.547 "raid_level": "raid0", 00:08:18.547 "superblock": false, 00:08:18.547 "num_base_bdevs": 3, 00:08:18.547 "num_base_bdevs_discovered": 3, 00:08:18.547 "num_base_bdevs_operational": 3, 00:08:18.547 "base_bdevs_list": [ 00:08:18.547 { 00:08:18.547 "name": "NewBaseBdev", 00:08:18.547 "uuid": "b5b8bd93-8893-47ba-8db7-1f8101924ec6", 00:08:18.547 "is_configured": true, 00:08:18.547 "data_offset": 0, 00:08:18.547 "data_size": 65536 00:08:18.547 }, 00:08:18.547 { 00:08:18.547 "name": "BaseBdev2", 00:08:18.547 "uuid": "208756dc-8205-49ce-b225-6205e0673803", 00:08:18.547 "is_configured": true, 00:08:18.547 "data_offset": 0, 00:08:18.547 "data_size": 65536 00:08:18.547 }, 00:08:18.547 { 00:08:18.547 "name": "BaseBdev3", 00:08:18.547 "uuid": "9d74845d-25c2-4fe6-bb22-364f8807162f", 00:08:18.547 "is_configured": true, 00:08:18.547 "data_offset": 0, 00:08:18.547 "data_size": 65536 00:08:18.547 } 00:08:18.547 ] 00:08:18.547 } 00:08:18.547 } 00:08:18.547 }' 00:08:18.547 15:12:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:18.547 15:12:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:08:18.547 BaseBdev2 00:08:18.547 BaseBdev3' 00:08:18.547 15:12:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:18.547 15:12:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:18.547 15:12:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:18.547 15:12:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:08:18.547 15:12:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.547 15:12:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.547 15:12:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:18.547 15:12:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.547 15:12:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:18.547 15:12:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:18.547 15:12:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:18.547 15:12:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:18.547 15:12:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.547 15:12:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.547 15:12:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:18.547 15:12:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.547 15:12:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:18.547 15:12:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:18.548 15:12:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:18.548 15:12:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:18.548 15:12:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:18.548 15:12:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.548 15:12:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.548 15:12:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.806 15:12:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:18.806 15:12:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:18.806 15:12:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:18.806 15:12:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.806 15:12:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.806 [2024-11-27 15:12:46.675832] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:18.806 [2024-11-27 15:12:46.675863] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:18.806 [2024-11-27 15:12:46.675947] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:18.806 [2024-11-27 15:12:46.676000] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:18.806 [2024-11-27 15:12:46.676011] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name Existed_Raid, state offline 00:08:18.806 15:12:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.806 15:12:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 75195 00:08:18.806 15:12:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 75195 ']' 00:08:18.807 15:12:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 75195 00:08:18.807 15:12:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:08:18.807 15:12:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:18.807 15:12:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 75195 00:08:18.807 15:12:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:18.807 15:12:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:18.807 15:12:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 75195' 00:08:18.807 killing process with pid 75195 00:08:18.807 15:12:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 75195 00:08:18.807 [2024-11-27 15:12:46.728071] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:18.807 15:12:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 75195 00:08:18.807 [2024-11-27 15:12:46.759958] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:19.066 15:12:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:08:19.066 00:08:19.066 real 0m9.093s 00:08:19.066 user 0m15.513s 00:08:19.066 sys 0m1.878s 00:08:19.066 15:12:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:19.066 15:12:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.066 ************************************ 00:08:19.066 END TEST raid_state_function_test 00:08:19.066 ************************************ 00:08:19.066 15:12:47 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 3 true 00:08:19.066 15:12:47 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:19.066 15:12:47 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:19.066 15:12:47 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:19.066 ************************************ 00:08:19.066 START TEST raid_state_function_test_sb 00:08:19.066 ************************************ 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 3 true 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=75800 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 75800' 00:08:19.066 Process raid pid: 75800 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 75800 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 75800 ']' 00:08:19.066 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:19.066 15:12:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:19.066 [2024-11-27 15:12:47.149198] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:08:19.066 [2024-11-27 15:12:47.149322] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:19.325 [2024-11-27 15:12:47.321067] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:19.326 [2024-11-27 15:12:47.347974] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:19.326 [2024-11-27 15:12:47.391652] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:19.326 [2024-11-27 15:12:47.391698] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:19.894 15:12:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:19.894 15:12:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:08:19.894 15:12:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:19.894 15:12:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.894 15:12:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:19.894 [2024-11-27 15:12:47.995970] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:19.894 [2024-11-27 15:12:47.996073] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:19.894 [2024-11-27 15:12:47.996105] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:19.894 [2024-11-27 15:12:47.996129] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:19.894 [2024-11-27 15:12:47.996166] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:19.894 [2024-11-27 15:12:47.996190] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:20.153 15:12:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.153 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:20.153 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:20.153 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:20.153 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:20.153 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:20.153 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:20.153 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:20.153 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:20.153 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:20.153 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:20.153 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:20.153 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:20.153 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.153 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:20.153 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.153 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:20.153 "name": "Existed_Raid", 00:08:20.153 "uuid": "fb00495f-12d8-4711-be5f-6869c6ae75a6", 00:08:20.153 "strip_size_kb": 64, 00:08:20.153 "state": "configuring", 00:08:20.153 "raid_level": "raid0", 00:08:20.153 "superblock": true, 00:08:20.153 "num_base_bdevs": 3, 00:08:20.153 "num_base_bdevs_discovered": 0, 00:08:20.153 "num_base_bdevs_operational": 3, 00:08:20.153 "base_bdevs_list": [ 00:08:20.153 { 00:08:20.153 "name": "BaseBdev1", 00:08:20.153 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:20.153 "is_configured": false, 00:08:20.153 "data_offset": 0, 00:08:20.153 "data_size": 0 00:08:20.153 }, 00:08:20.153 { 00:08:20.153 "name": "BaseBdev2", 00:08:20.153 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:20.153 "is_configured": false, 00:08:20.153 "data_offset": 0, 00:08:20.153 "data_size": 0 00:08:20.153 }, 00:08:20.153 { 00:08:20.153 "name": "BaseBdev3", 00:08:20.153 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:20.153 "is_configured": false, 00:08:20.153 "data_offset": 0, 00:08:20.153 "data_size": 0 00:08:20.153 } 00:08:20.153 ] 00:08:20.153 }' 00:08:20.153 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:20.153 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:20.413 [2024-11-27 15:12:48.443866] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:20.413 [2024-11-27 15:12:48.443954] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:20.413 [2024-11-27 15:12:48.455881] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:20.413 [2024-11-27 15:12:48.455974] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:20.413 [2024-11-27 15:12:48.455988] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:20.413 [2024-11-27 15:12:48.455997] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:20.413 [2024-11-27 15:12:48.456005] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:20.413 [2024-11-27 15:12:48.456016] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:20.413 [2024-11-27 15:12:48.476910] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:20.413 BaseBdev1 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:20.413 [ 00:08:20.413 { 00:08:20.413 "name": "BaseBdev1", 00:08:20.413 "aliases": [ 00:08:20.413 "86bd4171-8750-40b4-a433-3d43ef2f4f64" 00:08:20.413 ], 00:08:20.413 "product_name": "Malloc disk", 00:08:20.413 "block_size": 512, 00:08:20.413 "num_blocks": 65536, 00:08:20.413 "uuid": "86bd4171-8750-40b4-a433-3d43ef2f4f64", 00:08:20.413 "assigned_rate_limits": { 00:08:20.413 "rw_ios_per_sec": 0, 00:08:20.413 "rw_mbytes_per_sec": 0, 00:08:20.413 "r_mbytes_per_sec": 0, 00:08:20.413 "w_mbytes_per_sec": 0 00:08:20.413 }, 00:08:20.413 "claimed": true, 00:08:20.413 "claim_type": "exclusive_write", 00:08:20.413 "zoned": false, 00:08:20.413 "supported_io_types": { 00:08:20.413 "read": true, 00:08:20.413 "write": true, 00:08:20.413 "unmap": true, 00:08:20.413 "flush": true, 00:08:20.413 "reset": true, 00:08:20.413 "nvme_admin": false, 00:08:20.413 "nvme_io": false, 00:08:20.413 "nvme_io_md": false, 00:08:20.413 "write_zeroes": true, 00:08:20.413 "zcopy": true, 00:08:20.413 "get_zone_info": false, 00:08:20.413 "zone_management": false, 00:08:20.413 "zone_append": false, 00:08:20.413 "compare": false, 00:08:20.413 "compare_and_write": false, 00:08:20.413 "abort": true, 00:08:20.413 "seek_hole": false, 00:08:20.413 "seek_data": false, 00:08:20.413 "copy": true, 00:08:20.413 "nvme_iov_md": false 00:08:20.413 }, 00:08:20.413 "memory_domains": [ 00:08:20.413 { 00:08:20.413 "dma_device_id": "system", 00:08:20.413 "dma_device_type": 1 00:08:20.413 }, 00:08:20.413 { 00:08:20.413 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:20.413 "dma_device_type": 2 00:08:20.413 } 00:08:20.413 ], 00:08:20.413 "driver_specific": {} 00:08:20.413 } 00:08:20.413 ] 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:20.413 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:20.673 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:20.673 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:20.673 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:20.673 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:20.673 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:20.673 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:20.673 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.673 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:20.673 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.673 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:20.673 "name": "Existed_Raid", 00:08:20.673 "uuid": "5840a627-da00-4d7a-b2dc-bb238c0f3984", 00:08:20.673 "strip_size_kb": 64, 00:08:20.673 "state": "configuring", 00:08:20.673 "raid_level": "raid0", 00:08:20.673 "superblock": true, 00:08:20.673 "num_base_bdevs": 3, 00:08:20.673 "num_base_bdevs_discovered": 1, 00:08:20.673 "num_base_bdevs_operational": 3, 00:08:20.673 "base_bdevs_list": [ 00:08:20.673 { 00:08:20.673 "name": "BaseBdev1", 00:08:20.673 "uuid": "86bd4171-8750-40b4-a433-3d43ef2f4f64", 00:08:20.673 "is_configured": true, 00:08:20.673 "data_offset": 2048, 00:08:20.673 "data_size": 63488 00:08:20.673 }, 00:08:20.673 { 00:08:20.673 "name": "BaseBdev2", 00:08:20.674 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:20.674 "is_configured": false, 00:08:20.674 "data_offset": 0, 00:08:20.674 "data_size": 0 00:08:20.674 }, 00:08:20.674 { 00:08:20.674 "name": "BaseBdev3", 00:08:20.674 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:20.674 "is_configured": false, 00:08:20.674 "data_offset": 0, 00:08:20.674 "data_size": 0 00:08:20.674 } 00:08:20.674 ] 00:08:20.674 }' 00:08:20.674 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:20.674 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:20.934 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:20.934 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.934 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:20.934 [2024-11-27 15:12:48.940178] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:20.934 [2024-11-27 15:12:48.940280] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:08:20.934 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.934 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:20.934 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.934 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:20.934 [2024-11-27 15:12:48.952197] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:20.934 [2024-11-27 15:12:48.954125] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:20.934 [2024-11-27 15:12:48.954203] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:20.934 [2024-11-27 15:12:48.954232] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:20.934 [2024-11-27 15:12:48.954256] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:20.934 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.934 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:20.934 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:20.934 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:20.934 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:20.934 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:20.934 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:20.934 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:20.934 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:20.934 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:20.934 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:20.934 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:20.934 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:20.934 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:20.934 15:12:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:20.934 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.934 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:20.934 15:12:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.934 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:20.934 "name": "Existed_Raid", 00:08:20.934 "uuid": "458b84f7-2f8b-4dab-918d-01150f669844", 00:08:20.934 "strip_size_kb": 64, 00:08:20.934 "state": "configuring", 00:08:20.934 "raid_level": "raid0", 00:08:20.934 "superblock": true, 00:08:20.934 "num_base_bdevs": 3, 00:08:20.934 "num_base_bdevs_discovered": 1, 00:08:20.934 "num_base_bdevs_operational": 3, 00:08:20.934 "base_bdevs_list": [ 00:08:20.934 { 00:08:20.934 "name": "BaseBdev1", 00:08:20.934 "uuid": "86bd4171-8750-40b4-a433-3d43ef2f4f64", 00:08:20.934 "is_configured": true, 00:08:20.934 "data_offset": 2048, 00:08:20.934 "data_size": 63488 00:08:20.934 }, 00:08:20.934 { 00:08:20.934 "name": "BaseBdev2", 00:08:20.934 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:20.934 "is_configured": false, 00:08:20.934 "data_offset": 0, 00:08:20.934 "data_size": 0 00:08:20.934 }, 00:08:20.934 { 00:08:20.934 "name": "BaseBdev3", 00:08:20.934 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:20.934 "is_configured": false, 00:08:20.934 "data_offset": 0, 00:08:20.934 "data_size": 0 00:08:20.934 } 00:08:20.934 ] 00:08:20.934 }' 00:08:20.934 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:20.934 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:21.502 [2024-11-27 15:12:49.410820] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:21.502 BaseBdev2 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:21.502 [ 00:08:21.502 { 00:08:21.502 "name": "BaseBdev2", 00:08:21.502 "aliases": [ 00:08:21.502 "22d5ea0a-0f0c-4033-acc2-5a105a484588" 00:08:21.502 ], 00:08:21.502 "product_name": "Malloc disk", 00:08:21.502 "block_size": 512, 00:08:21.502 "num_blocks": 65536, 00:08:21.502 "uuid": "22d5ea0a-0f0c-4033-acc2-5a105a484588", 00:08:21.502 "assigned_rate_limits": { 00:08:21.502 "rw_ios_per_sec": 0, 00:08:21.502 "rw_mbytes_per_sec": 0, 00:08:21.502 "r_mbytes_per_sec": 0, 00:08:21.502 "w_mbytes_per_sec": 0 00:08:21.502 }, 00:08:21.502 "claimed": true, 00:08:21.502 "claim_type": "exclusive_write", 00:08:21.502 "zoned": false, 00:08:21.502 "supported_io_types": { 00:08:21.502 "read": true, 00:08:21.502 "write": true, 00:08:21.502 "unmap": true, 00:08:21.502 "flush": true, 00:08:21.502 "reset": true, 00:08:21.502 "nvme_admin": false, 00:08:21.502 "nvme_io": false, 00:08:21.502 "nvme_io_md": false, 00:08:21.502 "write_zeroes": true, 00:08:21.502 "zcopy": true, 00:08:21.502 "get_zone_info": false, 00:08:21.502 "zone_management": false, 00:08:21.502 "zone_append": false, 00:08:21.502 "compare": false, 00:08:21.502 "compare_and_write": false, 00:08:21.502 "abort": true, 00:08:21.502 "seek_hole": false, 00:08:21.502 "seek_data": false, 00:08:21.502 "copy": true, 00:08:21.502 "nvme_iov_md": false 00:08:21.502 }, 00:08:21.502 "memory_domains": [ 00:08:21.502 { 00:08:21.502 "dma_device_id": "system", 00:08:21.502 "dma_device_type": 1 00:08:21.502 }, 00:08:21.502 { 00:08:21.502 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:21.502 "dma_device_type": 2 00:08:21.502 } 00:08:21.502 ], 00:08:21.502 "driver_specific": {} 00:08:21.502 } 00:08:21.502 ] 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.502 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:21.502 "name": "Existed_Raid", 00:08:21.502 "uuid": "458b84f7-2f8b-4dab-918d-01150f669844", 00:08:21.502 "strip_size_kb": 64, 00:08:21.502 "state": "configuring", 00:08:21.502 "raid_level": "raid0", 00:08:21.502 "superblock": true, 00:08:21.502 "num_base_bdevs": 3, 00:08:21.502 "num_base_bdevs_discovered": 2, 00:08:21.503 "num_base_bdevs_operational": 3, 00:08:21.503 "base_bdevs_list": [ 00:08:21.503 { 00:08:21.503 "name": "BaseBdev1", 00:08:21.503 "uuid": "86bd4171-8750-40b4-a433-3d43ef2f4f64", 00:08:21.503 "is_configured": true, 00:08:21.503 "data_offset": 2048, 00:08:21.503 "data_size": 63488 00:08:21.503 }, 00:08:21.503 { 00:08:21.503 "name": "BaseBdev2", 00:08:21.503 "uuid": "22d5ea0a-0f0c-4033-acc2-5a105a484588", 00:08:21.503 "is_configured": true, 00:08:21.503 "data_offset": 2048, 00:08:21.503 "data_size": 63488 00:08:21.503 }, 00:08:21.503 { 00:08:21.503 "name": "BaseBdev3", 00:08:21.503 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:21.503 "is_configured": false, 00:08:21.503 "data_offset": 0, 00:08:21.503 "data_size": 0 00:08:21.503 } 00:08:21.503 ] 00:08:21.503 }' 00:08:21.503 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:21.503 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:21.761 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:21.761 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.761 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:22.026 [2024-11-27 15:12:49.877056] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:22.026 [2024-11-27 15:12:49.877379] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:08:22.026 BaseBdev3 00:08:22.026 [2024-11-27 15:12:49.877452] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:22.026 [2024-11-27 15:12:49.877797] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:08:22.026 [2024-11-27 15:12:49.878020] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:08:22.026 [2024-11-27 15:12:49.878036] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:08:22.026 [2024-11-27 15:12:49.878177] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:22.026 [ 00:08:22.026 { 00:08:22.026 "name": "BaseBdev3", 00:08:22.026 "aliases": [ 00:08:22.026 "e7e44736-c54b-4971-ae40-9b4417f843c8" 00:08:22.026 ], 00:08:22.026 "product_name": "Malloc disk", 00:08:22.026 "block_size": 512, 00:08:22.026 "num_blocks": 65536, 00:08:22.026 "uuid": "e7e44736-c54b-4971-ae40-9b4417f843c8", 00:08:22.026 "assigned_rate_limits": { 00:08:22.026 "rw_ios_per_sec": 0, 00:08:22.026 "rw_mbytes_per_sec": 0, 00:08:22.026 "r_mbytes_per_sec": 0, 00:08:22.026 "w_mbytes_per_sec": 0 00:08:22.026 }, 00:08:22.026 "claimed": true, 00:08:22.026 "claim_type": "exclusive_write", 00:08:22.026 "zoned": false, 00:08:22.026 "supported_io_types": { 00:08:22.026 "read": true, 00:08:22.026 "write": true, 00:08:22.026 "unmap": true, 00:08:22.026 "flush": true, 00:08:22.026 "reset": true, 00:08:22.026 "nvme_admin": false, 00:08:22.026 "nvme_io": false, 00:08:22.026 "nvme_io_md": false, 00:08:22.026 "write_zeroes": true, 00:08:22.026 "zcopy": true, 00:08:22.026 "get_zone_info": false, 00:08:22.026 "zone_management": false, 00:08:22.026 "zone_append": false, 00:08:22.026 "compare": false, 00:08:22.026 "compare_and_write": false, 00:08:22.026 "abort": true, 00:08:22.026 "seek_hole": false, 00:08:22.026 "seek_data": false, 00:08:22.026 "copy": true, 00:08:22.026 "nvme_iov_md": false 00:08:22.026 }, 00:08:22.026 "memory_domains": [ 00:08:22.026 { 00:08:22.026 "dma_device_id": "system", 00:08:22.026 "dma_device_type": 1 00:08:22.026 }, 00:08:22.026 { 00:08:22.026 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:22.026 "dma_device_type": 2 00:08:22.026 } 00:08:22.026 ], 00:08:22.026 "driver_specific": {} 00:08:22.026 } 00:08:22.026 ] 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:22.026 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:22.026 "name": "Existed_Raid", 00:08:22.026 "uuid": "458b84f7-2f8b-4dab-918d-01150f669844", 00:08:22.026 "strip_size_kb": 64, 00:08:22.026 "state": "online", 00:08:22.026 "raid_level": "raid0", 00:08:22.026 "superblock": true, 00:08:22.026 "num_base_bdevs": 3, 00:08:22.026 "num_base_bdevs_discovered": 3, 00:08:22.026 "num_base_bdevs_operational": 3, 00:08:22.026 "base_bdevs_list": [ 00:08:22.026 { 00:08:22.026 "name": "BaseBdev1", 00:08:22.026 "uuid": "86bd4171-8750-40b4-a433-3d43ef2f4f64", 00:08:22.027 "is_configured": true, 00:08:22.027 "data_offset": 2048, 00:08:22.027 "data_size": 63488 00:08:22.027 }, 00:08:22.027 { 00:08:22.027 "name": "BaseBdev2", 00:08:22.027 "uuid": "22d5ea0a-0f0c-4033-acc2-5a105a484588", 00:08:22.027 "is_configured": true, 00:08:22.027 "data_offset": 2048, 00:08:22.027 "data_size": 63488 00:08:22.027 }, 00:08:22.027 { 00:08:22.027 "name": "BaseBdev3", 00:08:22.027 "uuid": "e7e44736-c54b-4971-ae40-9b4417f843c8", 00:08:22.027 "is_configured": true, 00:08:22.027 "data_offset": 2048, 00:08:22.027 "data_size": 63488 00:08:22.027 } 00:08:22.027 ] 00:08:22.027 }' 00:08:22.027 15:12:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:22.027 15:12:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:22.289 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:22.289 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:22.289 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:22.289 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:22.289 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:22.289 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:22.289 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:22.289 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:22.289 15:12:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:22.289 15:12:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:22.289 [2024-11-27 15:12:50.376595] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:22.548 15:12:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:22.548 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:22.548 "name": "Existed_Raid", 00:08:22.548 "aliases": [ 00:08:22.548 "458b84f7-2f8b-4dab-918d-01150f669844" 00:08:22.548 ], 00:08:22.548 "product_name": "Raid Volume", 00:08:22.548 "block_size": 512, 00:08:22.548 "num_blocks": 190464, 00:08:22.548 "uuid": "458b84f7-2f8b-4dab-918d-01150f669844", 00:08:22.548 "assigned_rate_limits": { 00:08:22.548 "rw_ios_per_sec": 0, 00:08:22.548 "rw_mbytes_per_sec": 0, 00:08:22.548 "r_mbytes_per_sec": 0, 00:08:22.548 "w_mbytes_per_sec": 0 00:08:22.548 }, 00:08:22.548 "claimed": false, 00:08:22.548 "zoned": false, 00:08:22.548 "supported_io_types": { 00:08:22.548 "read": true, 00:08:22.548 "write": true, 00:08:22.548 "unmap": true, 00:08:22.548 "flush": true, 00:08:22.548 "reset": true, 00:08:22.548 "nvme_admin": false, 00:08:22.548 "nvme_io": false, 00:08:22.548 "nvme_io_md": false, 00:08:22.548 "write_zeroes": true, 00:08:22.548 "zcopy": false, 00:08:22.548 "get_zone_info": false, 00:08:22.548 "zone_management": false, 00:08:22.548 "zone_append": false, 00:08:22.548 "compare": false, 00:08:22.548 "compare_and_write": false, 00:08:22.548 "abort": false, 00:08:22.548 "seek_hole": false, 00:08:22.548 "seek_data": false, 00:08:22.548 "copy": false, 00:08:22.548 "nvme_iov_md": false 00:08:22.548 }, 00:08:22.548 "memory_domains": [ 00:08:22.548 { 00:08:22.548 "dma_device_id": "system", 00:08:22.548 "dma_device_type": 1 00:08:22.548 }, 00:08:22.548 { 00:08:22.548 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:22.548 "dma_device_type": 2 00:08:22.548 }, 00:08:22.548 { 00:08:22.548 "dma_device_id": "system", 00:08:22.548 "dma_device_type": 1 00:08:22.548 }, 00:08:22.548 { 00:08:22.548 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:22.548 "dma_device_type": 2 00:08:22.548 }, 00:08:22.548 { 00:08:22.548 "dma_device_id": "system", 00:08:22.548 "dma_device_type": 1 00:08:22.548 }, 00:08:22.548 { 00:08:22.548 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:22.548 "dma_device_type": 2 00:08:22.548 } 00:08:22.548 ], 00:08:22.548 "driver_specific": { 00:08:22.548 "raid": { 00:08:22.548 "uuid": "458b84f7-2f8b-4dab-918d-01150f669844", 00:08:22.548 "strip_size_kb": 64, 00:08:22.548 "state": "online", 00:08:22.548 "raid_level": "raid0", 00:08:22.548 "superblock": true, 00:08:22.548 "num_base_bdevs": 3, 00:08:22.548 "num_base_bdevs_discovered": 3, 00:08:22.548 "num_base_bdevs_operational": 3, 00:08:22.548 "base_bdevs_list": [ 00:08:22.548 { 00:08:22.548 "name": "BaseBdev1", 00:08:22.548 "uuid": "86bd4171-8750-40b4-a433-3d43ef2f4f64", 00:08:22.548 "is_configured": true, 00:08:22.548 "data_offset": 2048, 00:08:22.548 "data_size": 63488 00:08:22.548 }, 00:08:22.548 { 00:08:22.548 "name": "BaseBdev2", 00:08:22.548 "uuid": "22d5ea0a-0f0c-4033-acc2-5a105a484588", 00:08:22.548 "is_configured": true, 00:08:22.548 "data_offset": 2048, 00:08:22.548 "data_size": 63488 00:08:22.548 }, 00:08:22.548 { 00:08:22.548 "name": "BaseBdev3", 00:08:22.548 "uuid": "e7e44736-c54b-4971-ae40-9b4417f843c8", 00:08:22.548 "is_configured": true, 00:08:22.548 "data_offset": 2048, 00:08:22.548 "data_size": 63488 00:08:22.548 } 00:08:22.548 ] 00:08:22.548 } 00:08:22.548 } 00:08:22.548 }' 00:08:22.548 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:22.548 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:22.548 BaseBdev2 00:08:22.548 BaseBdev3' 00:08:22.548 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:22.548 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:22.548 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:22.548 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:22.548 15:12:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:22.548 15:12:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:22.548 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:22.548 15:12:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:22.548 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:22.548 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:22.548 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:22.548 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:22.548 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:22.548 15:12:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:22.548 15:12:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:22.548 15:12:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:22.548 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:22.548 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:22.548 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:22.548 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:22.548 15:12:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:22.548 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:22.548 15:12:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:22.548 15:12:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:22.807 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:22.807 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:22.807 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:22.807 15:12:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:22.807 15:12:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:22.807 [2024-11-27 15:12:50.659923] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:22.807 [2024-11-27 15:12:50.659996] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:22.807 [2024-11-27 15:12:50.660088] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:22.807 15:12:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:22.807 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:22.807 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:08:22.807 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:22.807 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:08:22.807 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:08:22.807 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 2 00:08:22.807 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:22.807 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:08:22.807 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:22.807 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:22.807 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:22.807 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:22.807 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:22.807 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:22.807 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:22.807 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:22.807 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:22.807 15:12:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:22.807 15:12:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:22.807 15:12:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:22.807 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:22.807 "name": "Existed_Raid", 00:08:22.807 "uuid": "458b84f7-2f8b-4dab-918d-01150f669844", 00:08:22.807 "strip_size_kb": 64, 00:08:22.807 "state": "offline", 00:08:22.807 "raid_level": "raid0", 00:08:22.807 "superblock": true, 00:08:22.807 "num_base_bdevs": 3, 00:08:22.807 "num_base_bdevs_discovered": 2, 00:08:22.807 "num_base_bdevs_operational": 2, 00:08:22.807 "base_bdevs_list": [ 00:08:22.807 { 00:08:22.807 "name": null, 00:08:22.807 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:22.807 "is_configured": false, 00:08:22.807 "data_offset": 0, 00:08:22.807 "data_size": 63488 00:08:22.807 }, 00:08:22.807 { 00:08:22.807 "name": "BaseBdev2", 00:08:22.807 "uuid": "22d5ea0a-0f0c-4033-acc2-5a105a484588", 00:08:22.807 "is_configured": true, 00:08:22.807 "data_offset": 2048, 00:08:22.807 "data_size": 63488 00:08:22.807 }, 00:08:22.807 { 00:08:22.807 "name": "BaseBdev3", 00:08:22.807 "uuid": "e7e44736-c54b-4971-ae40-9b4417f843c8", 00:08:22.807 "is_configured": true, 00:08:22.807 "data_offset": 2048, 00:08:22.807 "data_size": 63488 00:08:22.807 } 00:08:22.807 ] 00:08:22.807 }' 00:08:22.807 15:12:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:22.807 15:12:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.066 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:23.066 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:23.066 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:23.066 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:23.066 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.066 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.066 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.329 [2024-11-27 15:12:51.198702] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.329 [2024-11-27 15:12:51.266121] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:23.329 [2024-11-27 15:12:51.266214] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:23.329 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.330 BaseBdev2 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.330 [ 00:08:23.330 { 00:08:23.330 "name": "BaseBdev2", 00:08:23.330 "aliases": [ 00:08:23.330 "24eeb76f-6a3b-4544-a387-97e5d5dbc56c" 00:08:23.330 ], 00:08:23.330 "product_name": "Malloc disk", 00:08:23.330 "block_size": 512, 00:08:23.330 "num_blocks": 65536, 00:08:23.330 "uuid": "24eeb76f-6a3b-4544-a387-97e5d5dbc56c", 00:08:23.330 "assigned_rate_limits": { 00:08:23.330 "rw_ios_per_sec": 0, 00:08:23.330 "rw_mbytes_per_sec": 0, 00:08:23.330 "r_mbytes_per_sec": 0, 00:08:23.330 "w_mbytes_per_sec": 0 00:08:23.330 }, 00:08:23.330 "claimed": false, 00:08:23.330 "zoned": false, 00:08:23.330 "supported_io_types": { 00:08:23.330 "read": true, 00:08:23.330 "write": true, 00:08:23.330 "unmap": true, 00:08:23.330 "flush": true, 00:08:23.330 "reset": true, 00:08:23.330 "nvme_admin": false, 00:08:23.330 "nvme_io": false, 00:08:23.330 "nvme_io_md": false, 00:08:23.330 "write_zeroes": true, 00:08:23.330 "zcopy": true, 00:08:23.330 "get_zone_info": false, 00:08:23.330 "zone_management": false, 00:08:23.330 "zone_append": false, 00:08:23.330 "compare": false, 00:08:23.330 "compare_and_write": false, 00:08:23.330 "abort": true, 00:08:23.330 "seek_hole": false, 00:08:23.330 "seek_data": false, 00:08:23.330 "copy": true, 00:08:23.330 "nvme_iov_md": false 00:08:23.330 }, 00:08:23.330 "memory_domains": [ 00:08:23.330 { 00:08:23.330 "dma_device_id": "system", 00:08:23.330 "dma_device_type": 1 00:08:23.330 }, 00:08:23.330 { 00:08:23.330 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:23.330 "dma_device_type": 2 00:08:23.330 } 00:08:23.330 ], 00:08:23.330 "driver_specific": {} 00:08:23.330 } 00:08:23.330 ] 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.330 BaseBdev3 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.330 [ 00:08:23.330 { 00:08:23.330 "name": "BaseBdev3", 00:08:23.330 "aliases": [ 00:08:23.330 "4ca5263f-7ec1-4b23-8760-98a5063e1e0d" 00:08:23.330 ], 00:08:23.330 "product_name": "Malloc disk", 00:08:23.330 "block_size": 512, 00:08:23.330 "num_blocks": 65536, 00:08:23.330 "uuid": "4ca5263f-7ec1-4b23-8760-98a5063e1e0d", 00:08:23.330 "assigned_rate_limits": { 00:08:23.330 "rw_ios_per_sec": 0, 00:08:23.330 "rw_mbytes_per_sec": 0, 00:08:23.330 "r_mbytes_per_sec": 0, 00:08:23.330 "w_mbytes_per_sec": 0 00:08:23.330 }, 00:08:23.330 "claimed": false, 00:08:23.330 "zoned": false, 00:08:23.330 "supported_io_types": { 00:08:23.330 "read": true, 00:08:23.330 "write": true, 00:08:23.330 "unmap": true, 00:08:23.330 "flush": true, 00:08:23.330 "reset": true, 00:08:23.330 "nvme_admin": false, 00:08:23.330 "nvme_io": false, 00:08:23.330 "nvme_io_md": false, 00:08:23.330 "write_zeroes": true, 00:08:23.330 "zcopy": true, 00:08:23.330 "get_zone_info": false, 00:08:23.330 "zone_management": false, 00:08:23.330 "zone_append": false, 00:08:23.330 "compare": false, 00:08:23.330 "compare_and_write": false, 00:08:23.330 "abort": true, 00:08:23.330 "seek_hole": false, 00:08:23.330 "seek_data": false, 00:08:23.330 "copy": true, 00:08:23.330 "nvme_iov_md": false 00:08:23.330 }, 00:08:23.330 "memory_domains": [ 00:08:23.330 { 00:08:23.330 "dma_device_id": "system", 00:08:23.330 "dma_device_type": 1 00:08:23.330 }, 00:08:23.330 { 00:08:23.330 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:23.330 "dma_device_type": 2 00:08:23.330 } 00:08:23.330 ], 00:08:23.330 "driver_specific": {} 00:08:23.330 } 00:08:23.330 ] 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.330 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.590 [2024-11-27 15:12:51.438084] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:23.590 [2024-11-27 15:12:51.438175] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:23.590 [2024-11-27 15:12:51.438220] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:23.590 [2024-11-27 15:12:51.440085] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:23.590 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.590 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:23.590 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:23.590 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:23.590 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:23.590 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:23.590 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:23.590 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:23.590 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:23.590 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:23.590 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:23.590 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:23.590 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:23.590 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.590 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.590 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.590 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:23.590 "name": "Existed_Raid", 00:08:23.590 "uuid": "d5e89982-73fb-44e5-9611-2e29fb9bd896", 00:08:23.590 "strip_size_kb": 64, 00:08:23.590 "state": "configuring", 00:08:23.590 "raid_level": "raid0", 00:08:23.590 "superblock": true, 00:08:23.590 "num_base_bdevs": 3, 00:08:23.590 "num_base_bdevs_discovered": 2, 00:08:23.590 "num_base_bdevs_operational": 3, 00:08:23.590 "base_bdevs_list": [ 00:08:23.590 { 00:08:23.590 "name": "BaseBdev1", 00:08:23.590 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:23.590 "is_configured": false, 00:08:23.590 "data_offset": 0, 00:08:23.590 "data_size": 0 00:08:23.590 }, 00:08:23.590 { 00:08:23.590 "name": "BaseBdev2", 00:08:23.590 "uuid": "24eeb76f-6a3b-4544-a387-97e5d5dbc56c", 00:08:23.590 "is_configured": true, 00:08:23.590 "data_offset": 2048, 00:08:23.590 "data_size": 63488 00:08:23.590 }, 00:08:23.590 { 00:08:23.590 "name": "BaseBdev3", 00:08:23.590 "uuid": "4ca5263f-7ec1-4b23-8760-98a5063e1e0d", 00:08:23.590 "is_configured": true, 00:08:23.590 "data_offset": 2048, 00:08:23.590 "data_size": 63488 00:08:23.590 } 00:08:23.590 ] 00:08:23.590 }' 00:08:23.590 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:23.590 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.850 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:08:23.850 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.850 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.850 [2024-11-27 15:12:51.833431] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:23.850 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.850 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:23.850 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:23.850 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:23.850 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:23.850 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:23.850 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:23.850 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:23.850 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:23.850 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:23.850 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:23.850 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:23.850 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.850 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.850 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:23.850 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.850 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:23.850 "name": "Existed_Raid", 00:08:23.850 "uuid": "d5e89982-73fb-44e5-9611-2e29fb9bd896", 00:08:23.850 "strip_size_kb": 64, 00:08:23.850 "state": "configuring", 00:08:23.850 "raid_level": "raid0", 00:08:23.850 "superblock": true, 00:08:23.850 "num_base_bdevs": 3, 00:08:23.850 "num_base_bdevs_discovered": 1, 00:08:23.850 "num_base_bdevs_operational": 3, 00:08:23.850 "base_bdevs_list": [ 00:08:23.850 { 00:08:23.850 "name": "BaseBdev1", 00:08:23.850 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:23.850 "is_configured": false, 00:08:23.850 "data_offset": 0, 00:08:23.850 "data_size": 0 00:08:23.850 }, 00:08:23.850 { 00:08:23.850 "name": null, 00:08:23.850 "uuid": "24eeb76f-6a3b-4544-a387-97e5d5dbc56c", 00:08:23.850 "is_configured": false, 00:08:23.850 "data_offset": 0, 00:08:23.850 "data_size": 63488 00:08:23.850 }, 00:08:23.850 { 00:08:23.850 "name": "BaseBdev3", 00:08:23.850 "uuid": "4ca5263f-7ec1-4b23-8760-98a5063e1e0d", 00:08:23.850 "is_configured": true, 00:08:23.850 "data_offset": 2048, 00:08:23.850 "data_size": 63488 00:08:23.850 } 00:08:23.850 ] 00:08:23.850 }' 00:08:23.850 15:12:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:23.850 15:12:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.419 [2024-11-27 15:12:52.335737] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:24.419 BaseBdev1 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.419 [ 00:08:24.419 { 00:08:24.419 "name": "BaseBdev1", 00:08:24.419 "aliases": [ 00:08:24.419 "102ebc99-e297-4816-a88b-8b595baf93e3" 00:08:24.419 ], 00:08:24.419 "product_name": "Malloc disk", 00:08:24.419 "block_size": 512, 00:08:24.419 "num_blocks": 65536, 00:08:24.419 "uuid": "102ebc99-e297-4816-a88b-8b595baf93e3", 00:08:24.419 "assigned_rate_limits": { 00:08:24.419 "rw_ios_per_sec": 0, 00:08:24.419 "rw_mbytes_per_sec": 0, 00:08:24.419 "r_mbytes_per_sec": 0, 00:08:24.419 "w_mbytes_per_sec": 0 00:08:24.419 }, 00:08:24.419 "claimed": true, 00:08:24.419 "claim_type": "exclusive_write", 00:08:24.419 "zoned": false, 00:08:24.419 "supported_io_types": { 00:08:24.419 "read": true, 00:08:24.419 "write": true, 00:08:24.419 "unmap": true, 00:08:24.419 "flush": true, 00:08:24.419 "reset": true, 00:08:24.419 "nvme_admin": false, 00:08:24.419 "nvme_io": false, 00:08:24.419 "nvme_io_md": false, 00:08:24.419 "write_zeroes": true, 00:08:24.419 "zcopy": true, 00:08:24.419 "get_zone_info": false, 00:08:24.419 "zone_management": false, 00:08:24.419 "zone_append": false, 00:08:24.419 "compare": false, 00:08:24.419 "compare_and_write": false, 00:08:24.419 "abort": true, 00:08:24.419 "seek_hole": false, 00:08:24.419 "seek_data": false, 00:08:24.419 "copy": true, 00:08:24.419 "nvme_iov_md": false 00:08:24.419 }, 00:08:24.419 "memory_domains": [ 00:08:24.419 { 00:08:24.419 "dma_device_id": "system", 00:08:24.419 "dma_device_type": 1 00:08:24.419 }, 00:08:24.419 { 00:08:24.419 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:24.419 "dma_device_type": 2 00:08:24.419 } 00:08:24.419 ], 00:08:24.419 "driver_specific": {} 00:08:24.419 } 00:08:24.419 ] 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:24.419 "name": "Existed_Raid", 00:08:24.419 "uuid": "d5e89982-73fb-44e5-9611-2e29fb9bd896", 00:08:24.419 "strip_size_kb": 64, 00:08:24.419 "state": "configuring", 00:08:24.419 "raid_level": "raid0", 00:08:24.419 "superblock": true, 00:08:24.419 "num_base_bdevs": 3, 00:08:24.419 "num_base_bdevs_discovered": 2, 00:08:24.419 "num_base_bdevs_operational": 3, 00:08:24.419 "base_bdevs_list": [ 00:08:24.419 { 00:08:24.419 "name": "BaseBdev1", 00:08:24.419 "uuid": "102ebc99-e297-4816-a88b-8b595baf93e3", 00:08:24.419 "is_configured": true, 00:08:24.419 "data_offset": 2048, 00:08:24.419 "data_size": 63488 00:08:24.419 }, 00:08:24.419 { 00:08:24.419 "name": null, 00:08:24.419 "uuid": "24eeb76f-6a3b-4544-a387-97e5d5dbc56c", 00:08:24.419 "is_configured": false, 00:08:24.419 "data_offset": 0, 00:08:24.419 "data_size": 63488 00:08:24.419 }, 00:08:24.419 { 00:08:24.419 "name": "BaseBdev3", 00:08:24.419 "uuid": "4ca5263f-7ec1-4b23-8760-98a5063e1e0d", 00:08:24.419 "is_configured": true, 00:08:24.419 "data_offset": 2048, 00:08:24.419 "data_size": 63488 00:08:24.419 } 00:08:24.419 ] 00:08:24.419 }' 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:24.419 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.679 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:24.679 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:24.679 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.679 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:24.939 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:24.939 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:08:24.939 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:08:24.939 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:24.939 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.939 [2024-11-27 15:12:52.826974] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:24.939 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:24.939 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:24.939 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:24.939 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:24.939 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:24.939 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:24.939 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:24.939 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:24.939 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:24.939 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:24.939 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:24.939 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:24.939 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:24.939 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:24.939 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.939 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:24.939 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:24.939 "name": "Existed_Raid", 00:08:24.939 "uuid": "d5e89982-73fb-44e5-9611-2e29fb9bd896", 00:08:24.939 "strip_size_kb": 64, 00:08:24.939 "state": "configuring", 00:08:24.939 "raid_level": "raid0", 00:08:24.939 "superblock": true, 00:08:24.939 "num_base_bdevs": 3, 00:08:24.939 "num_base_bdevs_discovered": 1, 00:08:24.939 "num_base_bdevs_operational": 3, 00:08:24.939 "base_bdevs_list": [ 00:08:24.939 { 00:08:24.939 "name": "BaseBdev1", 00:08:24.939 "uuid": "102ebc99-e297-4816-a88b-8b595baf93e3", 00:08:24.939 "is_configured": true, 00:08:24.939 "data_offset": 2048, 00:08:24.939 "data_size": 63488 00:08:24.939 }, 00:08:24.939 { 00:08:24.939 "name": null, 00:08:24.939 "uuid": "24eeb76f-6a3b-4544-a387-97e5d5dbc56c", 00:08:24.939 "is_configured": false, 00:08:24.939 "data_offset": 0, 00:08:24.939 "data_size": 63488 00:08:24.939 }, 00:08:24.939 { 00:08:24.939 "name": null, 00:08:24.940 "uuid": "4ca5263f-7ec1-4b23-8760-98a5063e1e0d", 00:08:24.940 "is_configured": false, 00:08:24.940 "data_offset": 0, 00:08:24.940 "data_size": 63488 00:08:24.940 } 00:08:24.940 ] 00:08:24.940 }' 00:08:24.940 15:12:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:24.940 15:12:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.198 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:25.198 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:25.198 15:12:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.198 15:12:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.198 15:12:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.198 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:08:25.198 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:08:25.198 15:12:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.198 15:12:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.198 [2024-11-27 15:12:53.262194] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:25.198 15:12:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.198 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:25.198 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:25.198 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:25.198 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:25.198 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:25.198 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:25.198 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:25.198 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:25.198 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:25.198 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:25.199 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:25.199 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:25.199 15:12:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.199 15:12:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.199 15:12:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.457 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:25.457 "name": "Existed_Raid", 00:08:25.457 "uuid": "d5e89982-73fb-44e5-9611-2e29fb9bd896", 00:08:25.457 "strip_size_kb": 64, 00:08:25.457 "state": "configuring", 00:08:25.457 "raid_level": "raid0", 00:08:25.457 "superblock": true, 00:08:25.457 "num_base_bdevs": 3, 00:08:25.457 "num_base_bdevs_discovered": 2, 00:08:25.457 "num_base_bdevs_operational": 3, 00:08:25.457 "base_bdevs_list": [ 00:08:25.457 { 00:08:25.457 "name": "BaseBdev1", 00:08:25.457 "uuid": "102ebc99-e297-4816-a88b-8b595baf93e3", 00:08:25.457 "is_configured": true, 00:08:25.457 "data_offset": 2048, 00:08:25.457 "data_size": 63488 00:08:25.457 }, 00:08:25.457 { 00:08:25.457 "name": null, 00:08:25.457 "uuid": "24eeb76f-6a3b-4544-a387-97e5d5dbc56c", 00:08:25.457 "is_configured": false, 00:08:25.457 "data_offset": 0, 00:08:25.457 "data_size": 63488 00:08:25.457 }, 00:08:25.457 { 00:08:25.457 "name": "BaseBdev3", 00:08:25.457 "uuid": "4ca5263f-7ec1-4b23-8760-98a5063e1e0d", 00:08:25.457 "is_configured": true, 00:08:25.457 "data_offset": 2048, 00:08:25.457 "data_size": 63488 00:08:25.457 } 00:08:25.457 ] 00:08:25.457 }' 00:08:25.457 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:25.457 15:12:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.716 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:25.716 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:25.716 15:12:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.716 15:12:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.716 15:12:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.716 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:08:25.716 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:25.716 15:12:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.716 15:12:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.716 [2024-11-27 15:12:53.785351] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:25.716 15:12:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.716 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:25.716 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:25.716 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:25.716 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:25.716 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:25.716 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:25.716 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:25.716 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:25.716 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:25.716 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:25.716 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:25.716 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:25.716 15:12:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.716 15:12:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.976 15:12:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.976 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:25.976 "name": "Existed_Raid", 00:08:25.976 "uuid": "d5e89982-73fb-44e5-9611-2e29fb9bd896", 00:08:25.976 "strip_size_kb": 64, 00:08:25.976 "state": "configuring", 00:08:25.976 "raid_level": "raid0", 00:08:25.976 "superblock": true, 00:08:25.976 "num_base_bdevs": 3, 00:08:25.976 "num_base_bdevs_discovered": 1, 00:08:25.976 "num_base_bdevs_operational": 3, 00:08:25.976 "base_bdevs_list": [ 00:08:25.976 { 00:08:25.976 "name": null, 00:08:25.976 "uuid": "102ebc99-e297-4816-a88b-8b595baf93e3", 00:08:25.976 "is_configured": false, 00:08:25.976 "data_offset": 0, 00:08:25.976 "data_size": 63488 00:08:25.976 }, 00:08:25.976 { 00:08:25.976 "name": null, 00:08:25.976 "uuid": "24eeb76f-6a3b-4544-a387-97e5d5dbc56c", 00:08:25.976 "is_configured": false, 00:08:25.976 "data_offset": 0, 00:08:25.976 "data_size": 63488 00:08:25.976 }, 00:08:25.976 { 00:08:25.976 "name": "BaseBdev3", 00:08:25.976 "uuid": "4ca5263f-7ec1-4b23-8760-98a5063e1e0d", 00:08:25.976 "is_configured": true, 00:08:25.976 "data_offset": 2048, 00:08:25.976 "data_size": 63488 00:08:25.976 } 00:08:25.976 ] 00:08:25.976 }' 00:08:25.976 15:12:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:25.976 15:12:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.236 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:26.236 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:26.236 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.236 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.236 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.236 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:08:26.236 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:08:26.236 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.236 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.236 [2024-11-27 15:12:54.287876] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:26.236 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.236 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:26.236 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:26.236 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:26.236 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:26.236 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:26.236 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:26.236 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:26.236 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:26.236 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:26.236 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:26.236 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:26.236 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:26.236 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.236 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.236 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.495 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:26.495 "name": "Existed_Raid", 00:08:26.495 "uuid": "d5e89982-73fb-44e5-9611-2e29fb9bd896", 00:08:26.495 "strip_size_kb": 64, 00:08:26.495 "state": "configuring", 00:08:26.495 "raid_level": "raid0", 00:08:26.495 "superblock": true, 00:08:26.495 "num_base_bdevs": 3, 00:08:26.495 "num_base_bdevs_discovered": 2, 00:08:26.495 "num_base_bdevs_operational": 3, 00:08:26.495 "base_bdevs_list": [ 00:08:26.495 { 00:08:26.495 "name": null, 00:08:26.495 "uuid": "102ebc99-e297-4816-a88b-8b595baf93e3", 00:08:26.495 "is_configured": false, 00:08:26.495 "data_offset": 0, 00:08:26.495 "data_size": 63488 00:08:26.495 }, 00:08:26.495 { 00:08:26.495 "name": "BaseBdev2", 00:08:26.495 "uuid": "24eeb76f-6a3b-4544-a387-97e5d5dbc56c", 00:08:26.495 "is_configured": true, 00:08:26.495 "data_offset": 2048, 00:08:26.495 "data_size": 63488 00:08:26.495 }, 00:08:26.495 { 00:08:26.495 "name": "BaseBdev3", 00:08:26.495 "uuid": "4ca5263f-7ec1-4b23-8760-98a5063e1e0d", 00:08:26.495 "is_configured": true, 00:08:26.495 "data_offset": 2048, 00:08:26.495 "data_size": 63488 00:08:26.495 } 00:08:26.495 ] 00:08:26.495 }' 00:08:26.495 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:26.496 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.755 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:26.755 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.755 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.755 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:26.755 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.755 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:08:26.755 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:08:26.755 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:26.755 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.755 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.755 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.755 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 102ebc99-e297-4816-a88b-8b595baf93e3 00:08:26.755 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.755 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.755 [2024-11-27 15:12:54.826458] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:08:26.755 [2024-11-27 15:12:54.826683] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:08:26.755 [2024-11-27 15:12:54.826722] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:26.755 [2024-11-27 15:12:54.827031] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:08:26.755 NewBaseBdev 00:08:26.755 [2024-11-27 15:12:54.827193] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:08:26.755 [2024-11-27 15:12:54.827241] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006d00 00:08:26.755 [2024-11-27 15:12:54.827385] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:26.755 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.755 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:08:26.755 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:08:26.755 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:26.755 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:26.755 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:26.755 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:26.755 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:26.755 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.755 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.755 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.755 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:08:26.755 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.755 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.755 [ 00:08:26.755 { 00:08:26.755 "name": "NewBaseBdev", 00:08:26.755 "aliases": [ 00:08:26.755 "102ebc99-e297-4816-a88b-8b595baf93e3" 00:08:26.756 ], 00:08:26.756 "product_name": "Malloc disk", 00:08:26.756 "block_size": 512, 00:08:26.756 "num_blocks": 65536, 00:08:26.756 "uuid": "102ebc99-e297-4816-a88b-8b595baf93e3", 00:08:26.756 "assigned_rate_limits": { 00:08:26.756 "rw_ios_per_sec": 0, 00:08:26.756 "rw_mbytes_per_sec": 0, 00:08:26.756 "r_mbytes_per_sec": 0, 00:08:26.756 "w_mbytes_per_sec": 0 00:08:26.756 }, 00:08:26.756 "claimed": true, 00:08:26.756 "claim_type": "exclusive_write", 00:08:26.756 "zoned": false, 00:08:26.756 "supported_io_types": { 00:08:26.756 "read": true, 00:08:26.756 "write": true, 00:08:26.756 "unmap": true, 00:08:26.756 "flush": true, 00:08:26.756 "reset": true, 00:08:26.756 "nvme_admin": false, 00:08:26.756 "nvme_io": false, 00:08:26.756 "nvme_io_md": false, 00:08:26.756 "write_zeroes": true, 00:08:26.756 "zcopy": true, 00:08:26.756 "get_zone_info": false, 00:08:26.756 "zone_management": false, 00:08:26.756 "zone_append": false, 00:08:26.756 "compare": false, 00:08:26.756 "compare_and_write": false, 00:08:26.756 "abort": true, 00:08:26.756 "seek_hole": false, 00:08:26.756 "seek_data": false, 00:08:26.756 "copy": true, 00:08:26.756 "nvme_iov_md": false 00:08:27.015 }, 00:08:27.015 "memory_domains": [ 00:08:27.015 { 00:08:27.015 "dma_device_id": "system", 00:08:27.015 "dma_device_type": 1 00:08:27.015 }, 00:08:27.015 { 00:08:27.015 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:27.015 "dma_device_type": 2 00:08:27.015 } 00:08:27.015 ], 00:08:27.015 "driver_specific": {} 00:08:27.015 } 00:08:27.015 ] 00:08:27.015 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.015 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:27.015 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:08:27.015 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:27.015 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:27.015 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:27.015 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:27.015 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:27.015 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:27.015 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:27.015 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:27.015 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:27.015 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:27.015 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:27.015 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.015 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.015 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.015 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:27.015 "name": "Existed_Raid", 00:08:27.015 "uuid": "d5e89982-73fb-44e5-9611-2e29fb9bd896", 00:08:27.015 "strip_size_kb": 64, 00:08:27.015 "state": "online", 00:08:27.015 "raid_level": "raid0", 00:08:27.015 "superblock": true, 00:08:27.015 "num_base_bdevs": 3, 00:08:27.015 "num_base_bdevs_discovered": 3, 00:08:27.015 "num_base_bdevs_operational": 3, 00:08:27.015 "base_bdevs_list": [ 00:08:27.015 { 00:08:27.015 "name": "NewBaseBdev", 00:08:27.015 "uuid": "102ebc99-e297-4816-a88b-8b595baf93e3", 00:08:27.015 "is_configured": true, 00:08:27.015 "data_offset": 2048, 00:08:27.015 "data_size": 63488 00:08:27.015 }, 00:08:27.015 { 00:08:27.015 "name": "BaseBdev2", 00:08:27.015 "uuid": "24eeb76f-6a3b-4544-a387-97e5d5dbc56c", 00:08:27.015 "is_configured": true, 00:08:27.015 "data_offset": 2048, 00:08:27.015 "data_size": 63488 00:08:27.015 }, 00:08:27.015 { 00:08:27.015 "name": "BaseBdev3", 00:08:27.015 "uuid": "4ca5263f-7ec1-4b23-8760-98a5063e1e0d", 00:08:27.015 "is_configured": true, 00:08:27.015 "data_offset": 2048, 00:08:27.015 "data_size": 63488 00:08:27.015 } 00:08:27.015 ] 00:08:27.015 }' 00:08:27.015 15:12:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:27.015 15:12:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.275 15:12:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:08:27.275 15:12:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:27.275 15:12:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:27.275 15:12:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:27.275 15:12:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:27.275 15:12:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:27.275 15:12:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:27.275 15:12:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:27.275 15:12:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.275 15:12:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.275 [2024-11-27 15:12:55.333948] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:27.275 15:12:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.275 15:12:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:27.275 "name": "Existed_Raid", 00:08:27.275 "aliases": [ 00:08:27.275 "d5e89982-73fb-44e5-9611-2e29fb9bd896" 00:08:27.275 ], 00:08:27.275 "product_name": "Raid Volume", 00:08:27.275 "block_size": 512, 00:08:27.275 "num_blocks": 190464, 00:08:27.275 "uuid": "d5e89982-73fb-44e5-9611-2e29fb9bd896", 00:08:27.275 "assigned_rate_limits": { 00:08:27.275 "rw_ios_per_sec": 0, 00:08:27.275 "rw_mbytes_per_sec": 0, 00:08:27.275 "r_mbytes_per_sec": 0, 00:08:27.275 "w_mbytes_per_sec": 0 00:08:27.275 }, 00:08:27.275 "claimed": false, 00:08:27.275 "zoned": false, 00:08:27.275 "supported_io_types": { 00:08:27.275 "read": true, 00:08:27.275 "write": true, 00:08:27.275 "unmap": true, 00:08:27.275 "flush": true, 00:08:27.275 "reset": true, 00:08:27.275 "nvme_admin": false, 00:08:27.275 "nvme_io": false, 00:08:27.275 "nvme_io_md": false, 00:08:27.275 "write_zeroes": true, 00:08:27.275 "zcopy": false, 00:08:27.275 "get_zone_info": false, 00:08:27.275 "zone_management": false, 00:08:27.275 "zone_append": false, 00:08:27.275 "compare": false, 00:08:27.275 "compare_and_write": false, 00:08:27.275 "abort": false, 00:08:27.275 "seek_hole": false, 00:08:27.275 "seek_data": false, 00:08:27.275 "copy": false, 00:08:27.275 "nvme_iov_md": false 00:08:27.275 }, 00:08:27.275 "memory_domains": [ 00:08:27.275 { 00:08:27.275 "dma_device_id": "system", 00:08:27.275 "dma_device_type": 1 00:08:27.275 }, 00:08:27.275 { 00:08:27.275 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:27.275 "dma_device_type": 2 00:08:27.275 }, 00:08:27.275 { 00:08:27.275 "dma_device_id": "system", 00:08:27.275 "dma_device_type": 1 00:08:27.275 }, 00:08:27.275 { 00:08:27.275 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:27.275 "dma_device_type": 2 00:08:27.275 }, 00:08:27.275 { 00:08:27.275 "dma_device_id": "system", 00:08:27.275 "dma_device_type": 1 00:08:27.275 }, 00:08:27.275 { 00:08:27.275 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:27.275 "dma_device_type": 2 00:08:27.275 } 00:08:27.275 ], 00:08:27.275 "driver_specific": { 00:08:27.275 "raid": { 00:08:27.275 "uuid": "d5e89982-73fb-44e5-9611-2e29fb9bd896", 00:08:27.275 "strip_size_kb": 64, 00:08:27.275 "state": "online", 00:08:27.275 "raid_level": "raid0", 00:08:27.275 "superblock": true, 00:08:27.275 "num_base_bdevs": 3, 00:08:27.275 "num_base_bdevs_discovered": 3, 00:08:27.275 "num_base_bdevs_operational": 3, 00:08:27.275 "base_bdevs_list": [ 00:08:27.275 { 00:08:27.275 "name": "NewBaseBdev", 00:08:27.275 "uuid": "102ebc99-e297-4816-a88b-8b595baf93e3", 00:08:27.275 "is_configured": true, 00:08:27.275 "data_offset": 2048, 00:08:27.275 "data_size": 63488 00:08:27.275 }, 00:08:27.275 { 00:08:27.275 "name": "BaseBdev2", 00:08:27.275 "uuid": "24eeb76f-6a3b-4544-a387-97e5d5dbc56c", 00:08:27.275 "is_configured": true, 00:08:27.275 "data_offset": 2048, 00:08:27.275 "data_size": 63488 00:08:27.275 }, 00:08:27.275 { 00:08:27.275 "name": "BaseBdev3", 00:08:27.275 "uuid": "4ca5263f-7ec1-4b23-8760-98a5063e1e0d", 00:08:27.275 "is_configured": true, 00:08:27.275 "data_offset": 2048, 00:08:27.275 "data_size": 63488 00:08:27.275 } 00:08:27.275 ] 00:08:27.275 } 00:08:27.275 } 00:08:27.275 }' 00:08:27.275 15:12:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:08:27.534 BaseBdev2 00:08:27.534 BaseBdev3' 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.534 [2024-11-27 15:12:55.617108] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:27.534 [2024-11-27 15:12:55.617173] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:27.534 [2024-11-27 15:12:55.617268] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:27.534 [2024-11-27 15:12:55.617347] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:27.534 [2024-11-27 15:12:55.617380] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name Existed_Raid, state offline 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 75800 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 75800 ']' 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 75800 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:27.534 15:12:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 75800 00:08:27.793 killing process with pid 75800 00:08:27.793 15:12:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:27.793 15:12:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:27.793 15:12:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 75800' 00:08:27.793 15:12:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 75800 00:08:27.793 [2024-11-27 15:12:55.669743] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:27.793 15:12:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 75800 00:08:27.793 [2024-11-27 15:12:55.701540] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:28.053 15:12:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:08:28.053 00:08:28.053 real 0m8.860s 00:08:28.053 user 0m15.144s 00:08:28.053 sys 0m1.826s 00:08:28.053 ************************************ 00:08:28.053 END TEST raid_state_function_test_sb 00:08:28.053 ************************************ 00:08:28.053 15:12:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:28.053 15:12:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:28.053 15:12:55 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 3 00:08:28.053 15:12:55 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:08:28.053 15:12:55 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:28.053 15:12:55 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:28.053 ************************************ 00:08:28.053 START TEST raid_superblock_test 00:08:28.053 ************************************ 00:08:28.053 15:12:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid0 3 00:08:28.053 15:12:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:08:28.053 15:12:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:08:28.053 15:12:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:08:28.053 15:12:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:08:28.053 15:12:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:08:28.053 15:12:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:08:28.053 15:12:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:08:28.053 15:12:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:08:28.053 15:12:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:08:28.053 15:12:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:08:28.053 15:12:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:08:28.053 15:12:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:08:28.053 15:12:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:08:28.053 15:12:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:08:28.053 15:12:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:08:28.053 15:12:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:08:28.053 15:12:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=76409 00:08:28.053 15:12:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:08:28.053 15:12:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 76409 00:08:28.053 15:12:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 76409 ']' 00:08:28.053 15:12:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:28.053 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:28.053 15:12:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:28.053 15:12:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:28.053 15:12:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:28.053 15:12:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.053 [2024-11-27 15:12:56.072102] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:08:28.053 [2024-11-27 15:12:56.072273] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid76409 ] 00:08:28.312 [2024-11-27 15:12:56.242034] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:28.312 [2024-11-27 15:12:56.268011] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:28.312 [2024-11-27 15:12:56.311606] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:28.312 [2024-11-27 15:12:56.311746] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:28.880 15:12:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:28.880 15:12:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:08:28.880 15:12:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:08:28.880 15:12:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:28.880 15:12:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:08:28.880 15:12:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:08:28.880 15:12:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:08:28.880 15:12:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:28.880 15:12:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:28.880 15:12:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:28.880 15:12:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:08:28.880 15:12:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.880 15:12:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.880 malloc1 00:08:28.880 15:12:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.880 15:12:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:28.880 15:12:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.880 15:12:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.880 [2024-11-27 15:12:56.940524] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:28.880 [2024-11-27 15:12:56.940633] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:28.880 [2024-11-27 15:12:56.940671] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:08:28.880 [2024-11-27 15:12:56.940715] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:28.880 [2024-11-27 15:12:56.942828] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:28.880 [2024-11-27 15:12:56.942920] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:28.880 pt1 00:08:28.880 15:12:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.880 15:12:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:28.880 15:12:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:28.880 15:12:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:08:28.880 15:12:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:08:28.881 15:12:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:08:28.881 15:12:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:28.881 15:12:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:28.881 15:12:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:28.881 15:12:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:08:28.881 15:12:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.881 15:12:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.881 malloc2 00:08:28.881 15:12:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.881 15:12:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:28.881 15:12:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.881 15:12:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.881 [2024-11-27 15:12:56.969166] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:28.881 [2024-11-27 15:12:56.969259] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:28.881 [2024-11-27 15:12:56.969293] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:08:28.881 [2024-11-27 15:12:56.969319] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:28.881 [2024-11-27 15:12:56.971357] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:28.881 [2024-11-27 15:12:56.971438] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:28.881 pt2 00:08:28.881 15:12:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.881 15:12:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:28.881 15:12:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:28.881 15:12:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:08:28.881 15:12:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:08:28.881 15:12:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:08:28.881 15:12:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:28.881 15:12:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:28.881 15:12:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:28.881 15:12:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:08:28.881 15:12:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.881 15:12:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.141 malloc3 00:08:29.141 15:12:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.141 15:12:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:08:29.141 15:12:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.141 15:12:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.141 [2024-11-27 15:12:56.997873] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:08:29.141 [2024-11-27 15:12:56.997969] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:29.141 [2024-11-27 15:12:56.998005] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:08:29.141 [2024-11-27 15:12:56.998036] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:29.141 [2024-11-27 15:12:57.000120] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:29.141 [2024-11-27 15:12:57.000200] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:08:29.141 pt3 00:08:29.141 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.141 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:29.141 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:29.141 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:08:29.141 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.141 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.141 [2024-11-27 15:12:57.009903] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:29.141 [2024-11-27 15:12:57.011686] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:29.141 [2024-11-27 15:12:57.011825] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:08:29.141 [2024-11-27 15:12:57.012000] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:08:29.141 [2024-11-27 15:12:57.012053] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:29.141 [2024-11-27 15:12:57.012331] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:08:29.141 [2024-11-27 15:12:57.012506] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:08:29.141 [2024-11-27 15:12:57.012552] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:08:29.141 [2024-11-27 15:12:57.012714] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:29.141 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.141 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:29.141 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:29.141 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:29.141 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:29.141 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:29.141 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:29.141 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:29.141 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:29.141 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:29.141 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:29.141 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:29.141 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:29.141 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.141 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.141 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.141 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:29.141 "name": "raid_bdev1", 00:08:29.141 "uuid": "b29fd83c-d06f-4332-8d25-21987d3f0a0d", 00:08:29.141 "strip_size_kb": 64, 00:08:29.141 "state": "online", 00:08:29.141 "raid_level": "raid0", 00:08:29.141 "superblock": true, 00:08:29.141 "num_base_bdevs": 3, 00:08:29.141 "num_base_bdevs_discovered": 3, 00:08:29.141 "num_base_bdevs_operational": 3, 00:08:29.141 "base_bdevs_list": [ 00:08:29.141 { 00:08:29.141 "name": "pt1", 00:08:29.141 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:29.141 "is_configured": true, 00:08:29.141 "data_offset": 2048, 00:08:29.141 "data_size": 63488 00:08:29.141 }, 00:08:29.141 { 00:08:29.141 "name": "pt2", 00:08:29.141 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:29.141 "is_configured": true, 00:08:29.141 "data_offset": 2048, 00:08:29.141 "data_size": 63488 00:08:29.141 }, 00:08:29.141 { 00:08:29.141 "name": "pt3", 00:08:29.141 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:29.141 "is_configured": true, 00:08:29.141 "data_offset": 2048, 00:08:29.141 "data_size": 63488 00:08:29.141 } 00:08:29.141 ] 00:08:29.141 }' 00:08:29.141 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:29.141 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.401 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:08:29.401 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:29.401 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:29.401 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:29.401 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:29.401 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:29.401 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:29.401 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.401 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.401 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:29.401 [2024-11-27 15:12:57.441403] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:29.401 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.401 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:29.401 "name": "raid_bdev1", 00:08:29.401 "aliases": [ 00:08:29.401 "b29fd83c-d06f-4332-8d25-21987d3f0a0d" 00:08:29.401 ], 00:08:29.401 "product_name": "Raid Volume", 00:08:29.402 "block_size": 512, 00:08:29.402 "num_blocks": 190464, 00:08:29.402 "uuid": "b29fd83c-d06f-4332-8d25-21987d3f0a0d", 00:08:29.402 "assigned_rate_limits": { 00:08:29.402 "rw_ios_per_sec": 0, 00:08:29.402 "rw_mbytes_per_sec": 0, 00:08:29.402 "r_mbytes_per_sec": 0, 00:08:29.402 "w_mbytes_per_sec": 0 00:08:29.402 }, 00:08:29.402 "claimed": false, 00:08:29.402 "zoned": false, 00:08:29.402 "supported_io_types": { 00:08:29.402 "read": true, 00:08:29.402 "write": true, 00:08:29.402 "unmap": true, 00:08:29.402 "flush": true, 00:08:29.402 "reset": true, 00:08:29.402 "nvme_admin": false, 00:08:29.402 "nvme_io": false, 00:08:29.402 "nvme_io_md": false, 00:08:29.402 "write_zeroes": true, 00:08:29.402 "zcopy": false, 00:08:29.402 "get_zone_info": false, 00:08:29.402 "zone_management": false, 00:08:29.402 "zone_append": false, 00:08:29.402 "compare": false, 00:08:29.402 "compare_and_write": false, 00:08:29.402 "abort": false, 00:08:29.402 "seek_hole": false, 00:08:29.402 "seek_data": false, 00:08:29.402 "copy": false, 00:08:29.402 "nvme_iov_md": false 00:08:29.402 }, 00:08:29.402 "memory_domains": [ 00:08:29.402 { 00:08:29.402 "dma_device_id": "system", 00:08:29.402 "dma_device_type": 1 00:08:29.402 }, 00:08:29.402 { 00:08:29.402 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:29.402 "dma_device_type": 2 00:08:29.402 }, 00:08:29.402 { 00:08:29.402 "dma_device_id": "system", 00:08:29.402 "dma_device_type": 1 00:08:29.402 }, 00:08:29.402 { 00:08:29.402 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:29.402 "dma_device_type": 2 00:08:29.402 }, 00:08:29.402 { 00:08:29.402 "dma_device_id": "system", 00:08:29.402 "dma_device_type": 1 00:08:29.402 }, 00:08:29.402 { 00:08:29.402 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:29.402 "dma_device_type": 2 00:08:29.402 } 00:08:29.402 ], 00:08:29.402 "driver_specific": { 00:08:29.402 "raid": { 00:08:29.402 "uuid": "b29fd83c-d06f-4332-8d25-21987d3f0a0d", 00:08:29.402 "strip_size_kb": 64, 00:08:29.402 "state": "online", 00:08:29.402 "raid_level": "raid0", 00:08:29.402 "superblock": true, 00:08:29.402 "num_base_bdevs": 3, 00:08:29.402 "num_base_bdevs_discovered": 3, 00:08:29.402 "num_base_bdevs_operational": 3, 00:08:29.402 "base_bdevs_list": [ 00:08:29.402 { 00:08:29.402 "name": "pt1", 00:08:29.402 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:29.402 "is_configured": true, 00:08:29.402 "data_offset": 2048, 00:08:29.402 "data_size": 63488 00:08:29.402 }, 00:08:29.402 { 00:08:29.402 "name": "pt2", 00:08:29.402 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:29.402 "is_configured": true, 00:08:29.402 "data_offset": 2048, 00:08:29.402 "data_size": 63488 00:08:29.402 }, 00:08:29.402 { 00:08:29.402 "name": "pt3", 00:08:29.402 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:29.402 "is_configured": true, 00:08:29.402 "data_offset": 2048, 00:08:29.402 "data_size": 63488 00:08:29.402 } 00:08:29.402 ] 00:08:29.402 } 00:08:29.402 } 00:08:29.402 }' 00:08:29.402 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:29.661 pt2 00:08:29.661 pt3' 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.661 [2024-11-27 15:12:57.720856] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=b29fd83c-d06f-4332-8d25-21987d3f0a0d 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z b29fd83c-d06f-4332-8d25-21987d3f0a0d ']' 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.661 [2024-11-27 15:12:57.760533] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:29.661 [2024-11-27 15:12:57.760602] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:29.661 [2024-11-27 15:12:57.760703] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:29.661 [2024-11-27 15:12:57.760789] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:29.661 [2024-11-27 15:12:57.760840] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:08:29.661 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.920 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:29.920 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:08:29.920 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.920 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.920 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.920 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:08:29.920 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:08:29.920 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:29.920 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:08:29.920 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.920 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.920 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.920 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:29.920 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:08:29.920 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.920 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.920 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.920 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:29.920 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:08:29.920 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.921 [2024-11-27 15:12:57.912296] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:08:29.921 [2024-11-27 15:12:57.914118] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:08:29.921 [2024-11-27 15:12:57.914167] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:08:29.921 [2024-11-27 15:12:57.914216] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:08:29.921 [2024-11-27 15:12:57.914260] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:08:29.921 [2024-11-27 15:12:57.914278] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:08:29.921 [2024-11-27 15:12:57.914292] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:29.921 [2024-11-27 15:12:57.914303] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state configuring 00:08:29.921 request: 00:08:29.921 { 00:08:29.921 "name": "raid_bdev1", 00:08:29.921 "raid_level": "raid0", 00:08:29.921 "base_bdevs": [ 00:08:29.921 "malloc1", 00:08:29.921 "malloc2", 00:08:29.921 "malloc3" 00:08:29.921 ], 00:08:29.921 "strip_size_kb": 64, 00:08:29.921 "superblock": false, 00:08:29.921 "method": "bdev_raid_create", 00:08:29.921 "req_id": 1 00:08:29.921 } 00:08:29.921 Got JSON-RPC error response 00:08:29.921 response: 00:08:29.921 { 00:08:29.921 "code": -17, 00:08:29.921 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:08:29.921 } 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.921 [2024-11-27 15:12:57.976147] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:29.921 [2024-11-27 15:12:57.976240] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:29.921 [2024-11-27 15:12:57.976273] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:08:29.921 [2024-11-27 15:12:57.976306] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:29.921 [2024-11-27 15:12:57.978362] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:29.921 [2024-11-27 15:12:57.978436] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:29.921 [2024-11-27 15:12:57.978523] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:08:29.921 [2024-11-27 15:12:57.978576] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:29.921 pt1 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 3 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:29.921 15:12:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:30.181 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:30.181 "name": "raid_bdev1", 00:08:30.181 "uuid": "b29fd83c-d06f-4332-8d25-21987d3f0a0d", 00:08:30.181 "strip_size_kb": 64, 00:08:30.181 "state": "configuring", 00:08:30.181 "raid_level": "raid0", 00:08:30.181 "superblock": true, 00:08:30.181 "num_base_bdevs": 3, 00:08:30.181 "num_base_bdevs_discovered": 1, 00:08:30.181 "num_base_bdevs_operational": 3, 00:08:30.181 "base_bdevs_list": [ 00:08:30.181 { 00:08:30.181 "name": "pt1", 00:08:30.181 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:30.181 "is_configured": true, 00:08:30.181 "data_offset": 2048, 00:08:30.181 "data_size": 63488 00:08:30.181 }, 00:08:30.181 { 00:08:30.181 "name": null, 00:08:30.181 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:30.181 "is_configured": false, 00:08:30.181 "data_offset": 2048, 00:08:30.181 "data_size": 63488 00:08:30.181 }, 00:08:30.181 { 00:08:30.181 "name": null, 00:08:30.181 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:30.181 "is_configured": false, 00:08:30.181 "data_offset": 2048, 00:08:30.181 "data_size": 63488 00:08:30.181 } 00:08:30.181 ] 00:08:30.181 }' 00:08:30.181 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:30.181 15:12:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.441 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:08:30.441 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:30.441 15:12:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:30.441 15:12:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.441 [2024-11-27 15:12:58.471375] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:30.441 [2024-11-27 15:12:58.471500] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:30.441 [2024-11-27 15:12:58.471539] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:08:30.441 [2024-11-27 15:12:58.471573] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:30.441 [2024-11-27 15:12:58.472072] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:30.441 [2024-11-27 15:12:58.472141] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:30.441 [2024-11-27 15:12:58.472248] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:30.441 [2024-11-27 15:12:58.472312] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:30.441 pt2 00:08:30.441 15:12:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:30.441 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:08:30.441 15:12:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:30.441 15:12:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.441 [2024-11-27 15:12:58.483315] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:08:30.441 15:12:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:30.441 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 3 00:08:30.441 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:30.441 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:30.441 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:30.441 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:30.441 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:30.441 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:30.441 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:30.441 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:30.441 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:30.441 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:30.441 15:12:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:30.441 15:12:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.441 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:30.441 15:12:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:30.441 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:30.441 "name": "raid_bdev1", 00:08:30.441 "uuid": "b29fd83c-d06f-4332-8d25-21987d3f0a0d", 00:08:30.441 "strip_size_kb": 64, 00:08:30.441 "state": "configuring", 00:08:30.441 "raid_level": "raid0", 00:08:30.441 "superblock": true, 00:08:30.441 "num_base_bdevs": 3, 00:08:30.441 "num_base_bdevs_discovered": 1, 00:08:30.441 "num_base_bdevs_operational": 3, 00:08:30.441 "base_bdevs_list": [ 00:08:30.441 { 00:08:30.441 "name": "pt1", 00:08:30.441 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:30.441 "is_configured": true, 00:08:30.441 "data_offset": 2048, 00:08:30.441 "data_size": 63488 00:08:30.441 }, 00:08:30.441 { 00:08:30.441 "name": null, 00:08:30.441 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:30.441 "is_configured": false, 00:08:30.441 "data_offset": 0, 00:08:30.441 "data_size": 63488 00:08:30.441 }, 00:08:30.441 { 00:08:30.441 "name": null, 00:08:30.441 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:30.441 "is_configured": false, 00:08:30.441 "data_offset": 2048, 00:08:30.441 "data_size": 63488 00:08:30.441 } 00:08:30.441 ] 00:08:30.441 }' 00:08:30.441 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:30.441 15:12:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.011 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:08:31.011 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:31.011 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:31.011 15:12:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.011 15:12:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.011 [2024-11-27 15:12:58.934582] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:31.011 [2024-11-27 15:12:58.934705] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:31.011 [2024-11-27 15:12:58.934747] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:08:31.011 [2024-11-27 15:12:58.934775] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:31.011 [2024-11-27 15:12:58.935208] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:31.011 [2024-11-27 15:12:58.935270] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:31.011 [2024-11-27 15:12:58.935374] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:31.011 [2024-11-27 15:12:58.935423] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:31.011 pt2 00:08:31.011 15:12:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.011 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:31.011 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:31.011 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:08:31.011 15:12:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.011 15:12:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.011 [2024-11-27 15:12:58.946528] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:08:31.011 [2024-11-27 15:12:58.946615] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:31.011 [2024-11-27 15:12:58.946650] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:08:31.011 [2024-11-27 15:12:58.946677] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:31.011 [2024-11-27 15:12:58.947056] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:31.011 [2024-11-27 15:12:58.947112] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:08:31.011 [2024-11-27 15:12:58.947200] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:08:31.011 [2024-11-27 15:12:58.947255] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:08:31.011 [2024-11-27 15:12:58.947363] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:08:31.011 [2024-11-27 15:12:58.947399] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:31.011 [2024-11-27 15:12:58.947650] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:08:31.011 [2024-11-27 15:12:58.947834] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:08:31.011 [2024-11-27 15:12:58.947884] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:08:31.011 [2024-11-27 15:12:58.948041] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:31.011 pt3 00:08:31.011 15:12:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.011 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:31.011 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:31.011 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:31.012 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:31.012 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:31.012 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:31.012 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:31.012 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:31.012 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:31.012 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:31.012 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:31.012 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:31.012 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:31.012 15:12:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:31.012 15:12:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.012 15:12:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.012 15:12:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.012 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:31.012 "name": "raid_bdev1", 00:08:31.012 "uuid": "b29fd83c-d06f-4332-8d25-21987d3f0a0d", 00:08:31.012 "strip_size_kb": 64, 00:08:31.012 "state": "online", 00:08:31.012 "raid_level": "raid0", 00:08:31.012 "superblock": true, 00:08:31.012 "num_base_bdevs": 3, 00:08:31.012 "num_base_bdevs_discovered": 3, 00:08:31.012 "num_base_bdevs_operational": 3, 00:08:31.012 "base_bdevs_list": [ 00:08:31.012 { 00:08:31.012 "name": "pt1", 00:08:31.012 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:31.012 "is_configured": true, 00:08:31.012 "data_offset": 2048, 00:08:31.012 "data_size": 63488 00:08:31.012 }, 00:08:31.012 { 00:08:31.012 "name": "pt2", 00:08:31.012 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:31.012 "is_configured": true, 00:08:31.012 "data_offset": 2048, 00:08:31.012 "data_size": 63488 00:08:31.012 }, 00:08:31.012 { 00:08:31.012 "name": "pt3", 00:08:31.012 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:31.012 "is_configured": true, 00:08:31.012 "data_offset": 2048, 00:08:31.012 "data_size": 63488 00:08:31.012 } 00:08:31.012 ] 00:08:31.012 }' 00:08:31.012 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:31.012 15:12:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.270 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:08:31.270 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:31.270 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:31.270 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:31.270 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:31.270 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:31.529 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:31.529 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:31.529 15:12:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.529 15:12:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.529 [2024-11-27 15:12:59.382091] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:31.529 15:12:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.529 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:31.529 "name": "raid_bdev1", 00:08:31.529 "aliases": [ 00:08:31.529 "b29fd83c-d06f-4332-8d25-21987d3f0a0d" 00:08:31.529 ], 00:08:31.529 "product_name": "Raid Volume", 00:08:31.529 "block_size": 512, 00:08:31.529 "num_blocks": 190464, 00:08:31.529 "uuid": "b29fd83c-d06f-4332-8d25-21987d3f0a0d", 00:08:31.529 "assigned_rate_limits": { 00:08:31.529 "rw_ios_per_sec": 0, 00:08:31.529 "rw_mbytes_per_sec": 0, 00:08:31.529 "r_mbytes_per_sec": 0, 00:08:31.529 "w_mbytes_per_sec": 0 00:08:31.529 }, 00:08:31.529 "claimed": false, 00:08:31.529 "zoned": false, 00:08:31.529 "supported_io_types": { 00:08:31.529 "read": true, 00:08:31.529 "write": true, 00:08:31.529 "unmap": true, 00:08:31.529 "flush": true, 00:08:31.529 "reset": true, 00:08:31.529 "nvme_admin": false, 00:08:31.529 "nvme_io": false, 00:08:31.529 "nvme_io_md": false, 00:08:31.529 "write_zeroes": true, 00:08:31.529 "zcopy": false, 00:08:31.529 "get_zone_info": false, 00:08:31.529 "zone_management": false, 00:08:31.529 "zone_append": false, 00:08:31.529 "compare": false, 00:08:31.529 "compare_and_write": false, 00:08:31.529 "abort": false, 00:08:31.529 "seek_hole": false, 00:08:31.529 "seek_data": false, 00:08:31.529 "copy": false, 00:08:31.529 "nvme_iov_md": false 00:08:31.529 }, 00:08:31.529 "memory_domains": [ 00:08:31.529 { 00:08:31.529 "dma_device_id": "system", 00:08:31.529 "dma_device_type": 1 00:08:31.529 }, 00:08:31.529 { 00:08:31.529 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:31.529 "dma_device_type": 2 00:08:31.529 }, 00:08:31.529 { 00:08:31.529 "dma_device_id": "system", 00:08:31.529 "dma_device_type": 1 00:08:31.529 }, 00:08:31.529 { 00:08:31.529 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:31.529 "dma_device_type": 2 00:08:31.529 }, 00:08:31.529 { 00:08:31.529 "dma_device_id": "system", 00:08:31.529 "dma_device_type": 1 00:08:31.529 }, 00:08:31.529 { 00:08:31.529 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:31.530 "dma_device_type": 2 00:08:31.530 } 00:08:31.530 ], 00:08:31.530 "driver_specific": { 00:08:31.530 "raid": { 00:08:31.530 "uuid": "b29fd83c-d06f-4332-8d25-21987d3f0a0d", 00:08:31.530 "strip_size_kb": 64, 00:08:31.530 "state": "online", 00:08:31.530 "raid_level": "raid0", 00:08:31.530 "superblock": true, 00:08:31.530 "num_base_bdevs": 3, 00:08:31.530 "num_base_bdevs_discovered": 3, 00:08:31.530 "num_base_bdevs_operational": 3, 00:08:31.530 "base_bdevs_list": [ 00:08:31.530 { 00:08:31.530 "name": "pt1", 00:08:31.530 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:31.530 "is_configured": true, 00:08:31.530 "data_offset": 2048, 00:08:31.530 "data_size": 63488 00:08:31.530 }, 00:08:31.530 { 00:08:31.530 "name": "pt2", 00:08:31.530 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:31.530 "is_configured": true, 00:08:31.530 "data_offset": 2048, 00:08:31.530 "data_size": 63488 00:08:31.530 }, 00:08:31.530 { 00:08:31.530 "name": "pt3", 00:08:31.530 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:31.530 "is_configured": true, 00:08:31.530 "data_offset": 2048, 00:08:31.530 "data_size": 63488 00:08:31.530 } 00:08:31.530 ] 00:08:31.530 } 00:08:31.530 } 00:08:31.530 }' 00:08:31.530 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:31.530 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:31.530 pt2 00:08:31.530 pt3' 00:08:31.530 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:31.530 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:31.530 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:31.530 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:31.530 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:31.530 15:12:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.530 15:12:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.530 15:12:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.530 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:31.530 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:31.530 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:31.530 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:31.530 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:31.530 15:12:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.530 15:12:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.530 15:12:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.530 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:31.530 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:31.530 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:31.530 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:08:31.530 15:12:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.530 15:12:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.530 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:31.530 15:12:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.790 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:31.790 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:31.790 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:31.790 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:08:31.790 15:12:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.790 15:12:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.790 [2024-11-27 15:12:59.657523] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:31.790 15:12:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.790 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' b29fd83c-d06f-4332-8d25-21987d3f0a0d '!=' b29fd83c-d06f-4332-8d25-21987d3f0a0d ']' 00:08:31.790 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:08:31.790 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:31.790 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:31.790 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 76409 00:08:31.790 15:12:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 76409 ']' 00:08:31.790 15:12:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 76409 00:08:31.790 15:12:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:08:31.790 15:12:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:31.790 15:12:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 76409 00:08:31.790 killing process with pid 76409 00:08:31.790 15:12:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:31.790 15:12:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:31.790 15:12:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 76409' 00:08:31.790 15:12:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 76409 00:08:31.790 [2024-11-27 15:12:59.736155] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:31.790 [2024-11-27 15:12:59.736231] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:31.790 [2024-11-27 15:12:59.736293] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:31.790 [2024-11-27 15:12:59.736303] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:08:31.790 15:12:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 76409 00:08:31.790 [2024-11-27 15:12:59.769753] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:32.050 15:12:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:08:32.050 ************************************ 00:08:32.050 END TEST raid_superblock_test 00:08:32.050 ************************************ 00:08:32.050 00:08:32.050 real 0m4.009s 00:08:32.050 user 0m6.337s 00:08:32.050 sys 0m0.892s 00:08:32.050 15:12:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:32.050 15:12:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:32.050 15:13:00 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 3 read 00:08:32.050 15:13:00 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:32.050 15:13:00 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:32.050 15:13:00 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:32.050 ************************************ 00:08:32.050 START TEST raid_read_error_test 00:08:32.050 ************************************ 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 3 read 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.BPKFQ5VdDL 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=76651 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 76651 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 76651 ']' 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:32.050 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:32.050 15:13:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:32.310 [2024-11-27 15:13:00.163345] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:08:32.311 [2024-11-27 15:13:00.163632] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid76651 ] 00:08:32.311 [2024-11-27 15:13:00.334316] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:32.311 [2024-11-27 15:13:00.359580] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:32.311 [2024-11-27 15:13:00.403044] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:32.311 [2024-11-27 15:13:00.403083] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:33.250 15:13:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:33.250 15:13:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:08:33.250 15:13:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:33.250 15:13:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:33.250 15:13:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.250 15:13:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.250 BaseBdev1_malloc 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.250 true 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.250 [2024-11-27 15:13:01.027570] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:33.250 [2024-11-27 15:13:01.027667] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:33.250 [2024-11-27 15:13:01.027692] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:08:33.250 [2024-11-27 15:13:01.027723] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:33.250 [2024-11-27 15:13:01.029831] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:33.250 [2024-11-27 15:13:01.029871] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:33.250 BaseBdev1 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.250 BaseBdev2_malloc 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.250 true 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.250 [2024-11-27 15:13:01.068396] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:33.250 [2024-11-27 15:13:01.068496] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:33.250 [2024-11-27 15:13:01.068518] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:08:33.250 [2024-11-27 15:13:01.068527] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:33.250 [2024-11-27 15:13:01.070551] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:33.250 [2024-11-27 15:13:01.070590] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:33.250 BaseBdev2 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.250 BaseBdev3_malloc 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.250 true 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.250 [2024-11-27 15:13:01.109054] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:08:33.250 [2024-11-27 15:13:01.109140] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:33.250 [2024-11-27 15:13:01.109175] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:08:33.250 [2024-11-27 15:13:01.109203] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:33.250 [2024-11-27 15:13:01.111187] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:33.250 [2024-11-27 15:13:01.111256] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:08:33.250 BaseBdev3 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.250 15:13:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.250 [2024-11-27 15:13:01.121112] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:33.250 [2024-11-27 15:13:01.122853] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:33.250 [2024-11-27 15:13:01.122985] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:33.250 [2024-11-27 15:13:01.123159] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:08:33.250 [2024-11-27 15:13:01.123174] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:33.250 [2024-11-27 15:13:01.123403] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:08:33.250 [2024-11-27 15:13:01.123535] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:08:33.250 [2024-11-27 15:13:01.123545] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006d00 00:08:33.251 [2024-11-27 15:13:01.123668] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:33.251 15:13:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.251 15:13:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:33.251 15:13:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:33.251 15:13:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:33.251 15:13:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:33.251 15:13:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:33.251 15:13:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:33.251 15:13:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:33.251 15:13:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:33.251 15:13:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:33.251 15:13:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:33.251 15:13:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:33.251 15:13:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:33.251 15:13:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.251 15:13:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.251 15:13:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.251 15:13:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:33.251 "name": "raid_bdev1", 00:08:33.251 "uuid": "9f3d2d6a-4422-41f4-bf1e-db0aaf1d2e32", 00:08:33.251 "strip_size_kb": 64, 00:08:33.251 "state": "online", 00:08:33.251 "raid_level": "raid0", 00:08:33.251 "superblock": true, 00:08:33.251 "num_base_bdevs": 3, 00:08:33.251 "num_base_bdevs_discovered": 3, 00:08:33.251 "num_base_bdevs_operational": 3, 00:08:33.251 "base_bdevs_list": [ 00:08:33.251 { 00:08:33.251 "name": "BaseBdev1", 00:08:33.251 "uuid": "013cb546-b592-57eb-872e-ba619589d0de", 00:08:33.251 "is_configured": true, 00:08:33.251 "data_offset": 2048, 00:08:33.251 "data_size": 63488 00:08:33.251 }, 00:08:33.251 { 00:08:33.251 "name": "BaseBdev2", 00:08:33.251 "uuid": "22d0ba7d-2f67-50cf-be84-9822d0e618a2", 00:08:33.251 "is_configured": true, 00:08:33.251 "data_offset": 2048, 00:08:33.251 "data_size": 63488 00:08:33.251 }, 00:08:33.251 { 00:08:33.251 "name": "BaseBdev3", 00:08:33.251 "uuid": "758596ba-82e6-5c98-98cf-2e39268a3553", 00:08:33.251 "is_configured": true, 00:08:33.251 "data_offset": 2048, 00:08:33.251 "data_size": 63488 00:08:33.251 } 00:08:33.251 ] 00:08:33.251 }' 00:08:33.251 15:13:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:33.251 15:13:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.510 15:13:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:33.510 15:13:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:33.769 [2024-11-27 15:13:01.620607] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006560 00:08:34.708 15:13:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:08:34.708 15:13:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.708 15:13:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.708 15:13:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.708 15:13:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:34.708 15:13:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:08:34.708 15:13:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:08:34.708 15:13:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:34.708 15:13:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:34.708 15:13:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:34.708 15:13:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:34.708 15:13:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:34.708 15:13:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:34.708 15:13:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:34.708 15:13:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:34.708 15:13:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:34.708 15:13:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:34.708 15:13:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:34.708 15:13:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:34.708 15:13:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.708 15:13:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.708 15:13:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.708 15:13:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:34.708 "name": "raid_bdev1", 00:08:34.708 "uuid": "9f3d2d6a-4422-41f4-bf1e-db0aaf1d2e32", 00:08:34.708 "strip_size_kb": 64, 00:08:34.708 "state": "online", 00:08:34.708 "raid_level": "raid0", 00:08:34.708 "superblock": true, 00:08:34.708 "num_base_bdevs": 3, 00:08:34.708 "num_base_bdevs_discovered": 3, 00:08:34.708 "num_base_bdevs_operational": 3, 00:08:34.708 "base_bdevs_list": [ 00:08:34.708 { 00:08:34.708 "name": "BaseBdev1", 00:08:34.708 "uuid": "013cb546-b592-57eb-872e-ba619589d0de", 00:08:34.708 "is_configured": true, 00:08:34.708 "data_offset": 2048, 00:08:34.708 "data_size": 63488 00:08:34.708 }, 00:08:34.708 { 00:08:34.708 "name": "BaseBdev2", 00:08:34.708 "uuid": "22d0ba7d-2f67-50cf-be84-9822d0e618a2", 00:08:34.708 "is_configured": true, 00:08:34.708 "data_offset": 2048, 00:08:34.708 "data_size": 63488 00:08:34.708 }, 00:08:34.708 { 00:08:34.708 "name": "BaseBdev3", 00:08:34.708 "uuid": "758596ba-82e6-5c98-98cf-2e39268a3553", 00:08:34.708 "is_configured": true, 00:08:34.708 "data_offset": 2048, 00:08:34.708 "data_size": 63488 00:08:34.708 } 00:08:34.708 ] 00:08:34.708 }' 00:08:34.708 15:13:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:34.708 15:13:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.969 15:13:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:34.969 15:13:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.969 15:13:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.969 [2024-11-27 15:13:03.004298] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:34.969 [2024-11-27 15:13:03.004394] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:34.969 [2024-11-27 15:13:03.007044] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:34.969 [2024-11-27 15:13:03.007139] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:34.969 [2024-11-27 15:13:03.007196] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:34.969 [2024-11-27 15:13:03.007250] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name raid_bdev1, state offline 00:08:34.969 { 00:08:34.969 "results": [ 00:08:34.969 { 00:08:34.969 "job": "raid_bdev1", 00:08:34.969 "core_mask": "0x1", 00:08:34.969 "workload": "randrw", 00:08:34.969 "percentage": 50, 00:08:34.969 "status": "finished", 00:08:34.969 "queue_depth": 1, 00:08:34.969 "io_size": 131072, 00:08:34.969 "runtime": 1.384646, 00:08:34.969 "iops": 16638.18766673937, 00:08:34.969 "mibps": 2079.7734583424212, 00:08:34.969 "io_failed": 1, 00:08:34.969 "io_timeout": 0, 00:08:34.969 "avg_latency_us": 83.08816229780109, 00:08:34.969 "min_latency_us": 19.004366812227076, 00:08:34.969 "max_latency_us": 1337.907423580786 00:08:34.969 } 00:08:34.969 ], 00:08:34.969 "core_count": 1 00:08:34.969 } 00:08:34.969 15:13:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.969 15:13:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 76651 00:08:34.969 15:13:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 76651 ']' 00:08:34.969 15:13:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 76651 00:08:34.969 15:13:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:08:34.969 15:13:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:34.969 15:13:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 76651 00:08:34.969 15:13:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:34.969 15:13:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:34.969 15:13:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 76651' 00:08:34.969 killing process with pid 76651 00:08:34.969 15:13:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 76651 00:08:34.969 [2024-11-27 15:13:03.051564] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:34.969 15:13:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 76651 00:08:35.252 [2024-11-27 15:13:03.077918] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:35.252 15:13:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:35.252 15:13:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.BPKFQ5VdDL 00:08:35.252 15:13:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:35.252 15:13:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.72 00:08:35.252 15:13:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:08:35.252 15:13:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:35.252 15:13:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:35.252 15:13:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.72 != \0\.\0\0 ]] 00:08:35.252 00:08:35.252 real 0m3.243s 00:08:35.252 user 0m4.085s 00:08:35.252 sys 0m0.553s 00:08:35.252 15:13:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:35.252 ************************************ 00:08:35.252 END TEST raid_read_error_test 00:08:35.252 ************************************ 00:08:35.252 15:13:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.523 15:13:03 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 3 write 00:08:35.523 15:13:03 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:35.523 15:13:03 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:35.523 15:13:03 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:35.523 ************************************ 00:08:35.523 START TEST raid_write_error_test 00:08:35.523 ************************************ 00:08:35.523 15:13:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 3 write 00:08:35.523 15:13:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:08:35.523 15:13:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:08:35.523 15:13:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:08:35.523 15:13:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:35.524 15:13:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:35.524 15:13:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:35.524 15:13:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:35.524 15:13:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:35.524 15:13:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:35.524 15:13:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:35.524 15:13:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:35.524 15:13:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:08:35.524 15:13:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:35.524 15:13:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:35.524 15:13:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:35.524 15:13:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:35.524 15:13:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:35.524 15:13:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:35.524 15:13:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:35.524 15:13:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:35.524 15:13:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:35.524 15:13:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:08:35.524 15:13:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:08:35.524 15:13:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:08:35.524 15:13:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:35.524 15:13:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.ldnyR3KMvc 00:08:35.524 15:13:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=76780 00:08:35.524 15:13:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:35.524 15:13:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 76780 00:08:35.524 15:13:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 76780 ']' 00:08:35.524 15:13:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:35.524 15:13:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:35.524 15:13:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:35.524 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:35.524 15:13:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:35.524 15:13:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.524 [2024-11-27 15:13:03.485699] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:08:35.524 [2024-11-27 15:13:03.485961] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid76780 ] 00:08:35.784 [2024-11-27 15:13:03.663432] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:35.784 [2024-11-27 15:13:03.690312] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:35.784 [2024-11-27 15:13:03.733779] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:35.784 [2024-11-27 15:13:03.733910] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.353 BaseBdev1_malloc 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.353 true 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.353 [2024-11-27 15:13:04.350223] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:36.353 [2024-11-27 15:13:04.350316] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:36.353 [2024-11-27 15:13:04.350362] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:08:36.353 [2024-11-27 15:13:04.350391] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:36.353 [2024-11-27 15:13:04.352548] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:36.353 [2024-11-27 15:13:04.352624] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:36.353 BaseBdev1 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.353 BaseBdev2_malloc 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.353 true 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.353 [2024-11-27 15:13:04.390935] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:36.353 [2024-11-27 15:13:04.391022] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:36.353 [2024-11-27 15:13:04.391057] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:08:36.353 [2024-11-27 15:13:04.391084] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:36.353 [2024-11-27 15:13:04.393149] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:36.353 [2024-11-27 15:13:04.393221] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:36.353 BaseBdev2 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.353 BaseBdev3_malloc 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.353 true 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.353 [2024-11-27 15:13:04.431588] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:08:36.353 [2024-11-27 15:13:04.431673] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:36.353 [2024-11-27 15:13:04.431720] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:08:36.353 [2024-11-27 15:13:04.431754] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:36.353 [2024-11-27 15:13:04.433772] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:36.353 [2024-11-27 15:13:04.433843] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:08:36.353 BaseBdev3 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.353 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.353 [2024-11-27 15:13:04.443624] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:36.353 [2024-11-27 15:13:04.445434] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:36.353 [2024-11-27 15:13:04.445550] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:36.353 [2024-11-27 15:13:04.445738] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:08:36.353 [2024-11-27 15:13:04.445787] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:36.353 [2024-11-27 15:13:04.446051] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:08:36.353 [2024-11-27 15:13:04.446218] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:08:36.354 [2024-11-27 15:13:04.446260] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006d00 00:08:36.354 [2024-11-27 15:13:04.446434] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:36.354 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.354 15:13:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:36.354 15:13:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:36.354 15:13:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:36.354 15:13:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:36.354 15:13:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:36.354 15:13:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:36.354 15:13:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:36.354 15:13:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:36.354 15:13:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:36.354 15:13:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:36.354 15:13:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:36.354 15:13:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:36.354 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.354 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.614 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.614 15:13:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:36.614 "name": "raid_bdev1", 00:08:36.614 "uuid": "6faee0f1-58e4-465d-a004-538653d1bf1a", 00:08:36.614 "strip_size_kb": 64, 00:08:36.614 "state": "online", 00:08:36.614 "raid_level": "raid0", 00:08:36.614 "superblock": true, 00:08:36.614 "num_base_bdevs": 3, 00:08:36.614 "num_base_bdevs_discovered": 3, 00:08:36.614 "num_base_bdevs_operational": 3, 00:08:36.614 "base_bdevs_list": [ 00:08:36.614 { 00:08:36.614 "name": "BaseBdev1", 00:08:36.614 "uuid": "bd19d748-43b9-53bd-ac2c-2ffa5014bf22", 00:08:36.614 "is_configured": true, 00:08:36.614 "data_offset": 2048, 00:08:36.614 "data_size": 63488 00:08:36.614 }, 00:08:36.614 { 00:08:36.614 "name": "BaseBdev2", 00:08:36.614 "uuid": "9cbf91a2-0d97-56b8-b5dc-7ca6e9024244", 00:08:36.614 "is_configured": true, 00:08:36.614 "data_offset": 2048, 00:08:36.614 "data_size": 63488 00:08:36.614 }, 00:08:36.614 { 00:08:36.614 "name": "BaseBdev3", 00:08:36.614 "uuid": "8d67da6f-c928-5a1d-b410-69070ada3647", 00:08:36.614 "is_configured": true, 00:08:36.614 "data_offset": 2048, 00:08:36.614 "data_size": 63488 00:08:36.614 } 00:08:36.614 ] 00:08:36.614 }' 00:08:36.614 15:13:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:36.614 15:13:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.873 15:13:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:36.873 15:13:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:37.132 [2024-11-27 15:13:04.979111] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006560 00:08:38.071 15:13:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:08:38.071 15:13:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:38.071 15:13:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.071 15:13:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:38.071 15:13:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:38.071 15:13:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:08:38.071 15:13:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:08:38.071 15:13:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:38.071 15:13:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:38.071 15:13:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:38.071 15:13:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:38.071 15:13:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:38.071 15:13:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:38.071 15:13:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:38.071 15:13:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:38.071 15:13:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:38.071 15:13:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:38.071 15:13:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:38.071 15:13:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:38.071 15:13:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:38.071 15:13:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.071 15:13:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:38.071 15:13:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:38.071 "name": "raid_bdev1", 00:08:38.071 "uuid": "6faee0f1-58e4-465d-a004-538653d1bf1a", 00:08:38.071 "strip_size_kb": 64, 00:08:38.071 "state": "online", 00:08:38.071 "raid_level": "raid0", 00:08:38.071 "superblock": true, 00:08:38.071 "num_base_bdevs": 3, 00:08:38.071 "num_base_bdevs_discovered": 3, 00:08:38.071 "num_base_bdevs_operational": 3, 00:08:38.071 "base_bdevs_list": [ 00:08:38.071 { 00:08:38.071 "name": "BaseBdev1", 00:08:38.071 "uuid": "bd19d748-43b9-53bd-ac2c-2ffa5014bf22", 00:08:38.071 "is_configured": true, 00:08:38.071 "data_offset": 2048, 00:08:38.071 "data_size": 63488 00:08:38.071 }, 00:08:38.071 { 00:08:38.071 "name": "BaseBdev2", 00:08:38.071 "uuid": "9cbf91a2-0d97-56b8-b5dc-7ca6e9024244", 00:08:38.071 "is_configured": true, 00:08:38.071 "data_offset": 2048, 00:08:38.071 "data_size": 63488 00:08:38.071 }, 00:08:38.071 { 00:08:38.071 "name": "BaseBdev3", 00:08:38.071 "uuid": "8d67da6f-c928-5a1d-b410-69070ada3647", 00:08:38.071 "is_configured": true, 00:08:38.071 "data_offset": 2048, 00:08:38.071 "data_size": 63488 00:08:38.071 } 00:08:38.071 ] 00:08:38.071 }' 00:08:38.071 15:13:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:38.071 15:13:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.331 15:13:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:38.331 15:13:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:38.331 15:13:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.331 [2024-11-27 15:13:06.379118] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:38.331 [2024-11-27 15:13:06.379197] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:38.331 [2024-11-27 15:13:06.381734] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:38.331 [2024-11-27 15:13:06.381828] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:38.331 [2024-11-27 15:13:06.381881] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:38.331 [2024-11-27 15:13:06.381962] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name raid_bdev1, state offline 00:08:38.331 { 00:08:38.331 "results": [ 00:08:38.331 { 00:08:38.331 "job": "raid_bdev1", 00:08:38.331 "core_mask": "0x1", 00:08:38.331 "workload": "randrw", 00:08:38.331 "percentage": 50, 00:08:38.331 "status": "finished", 00:08:38.331 "queue_depth": 1, 00:08:38.331 "io_size": 131072, 00:08:38.331 "runtime": 1.400882, 00:08:38.331 "iops": 16631.66490825066, 00:08:38.331 "mibps": 2078.9581135313324, 00:08:38.331 "io_failed": 1, 00:08:38.331 "io_timeout": 0, 00:08:38.331 "avg_latency_us": 83.0449400078715, 00:08:38.331 "min_latency_us": 23.811353711790392, 00:08:38.331 "max_latency_us": 1380.8349344978167 00:08:38.331 } 00:08:38.331 ], 00:08:38.331 "core_count": 1 00:08:38.331 } 00:08:38.331 15:13:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:38.331 15:13:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 76780 00:08:38.331 15:13:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 76780 ']' 00:08:38.331 15:13:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 76780 00:08:38.331 15:13:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:08:38.331 15:13:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:38.331 15:13:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 76780 00:08:38.331 15:13:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:38.331 15:13:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:38.331 15:13:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 76780' 00:08:38.331 killing process with pid 76780 00:08:38.331 15:13:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 76780 00:08:38.331 [2024-11-27 15:13:06.429743] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:38.331 15:13:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 76780 00:08:38.590 [2024-11-27 15:13:06.455607] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:38.590 15:13:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.ldnyR3KMvc 00:08:38.590 15:13:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:38.590 15:13:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:38.590 15:13:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.71 00:08:38.590 15:13:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:08:38.590 15:13:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:38.590 15:13:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:38.590 15:13:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.71 != \0\.\0\0 ]] 00:08:38.590 00:08:38.590 real 0m3.303s 00:08:38.590 user 0m4.203s 00:08:38.590 sys 0m0.547s 00:08:38.591 15:13:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:38.591 ************************************ 00:08:38.591 END TEST raid_write_error_test 00:08:38.591 ************************************ 00:08:38.591 15:13:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.850 15:13:06 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:08:38.850 15:13:06 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 3 false 00:08:38.850 15:13:06 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:38.850 15:13:06 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:38.851 15:13:06 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:38.851 ************************************ 00:08:38.851 START TEST raid_state_function_test 00:08:38.851 ************************************ 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 3 false 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=76907 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:38.851 Process raid pid: 76907 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 76907' 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 76907 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 76907 ']' 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:38.851 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:38.851 15:13:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.851 [2024-11-27 15:13:06.850635] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:08:38.851 [2024-11-27 15:13:06.850874] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:39.110 [2024-11-27 15:13:07.018986] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:39.110 [2024-11-27 15:13:07.044747] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:39.110 [2024-11-27 15:13:07.087822] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:39.110 [2024-11-27 15:13:07.087990] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:39.678 15:13:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:39.678 15:13:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:08:39.678 15:13:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:39.678 15:13:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:39.678 15:13:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.678 [2024-11-27 15:13:07.671886] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:39.678 [2024-11-27 15:13:07.671989] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:39.678 [2024-11-27 15:13:07.672019] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:39.678 [2024-11-27 15:13:07.672042] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:39.678 [2024-11-27 15:13:07.672062] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:39.678 [2024-11-27 15:13:07.672087] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:39.678 15:13:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:39.678 15:13:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:39.678 15:13:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:39.678 15:13:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:39.678 15:13:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:39.678 15:13:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:39.678 15:13:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:39.678 15:13:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:39.678 15:13:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:39.678 15:13:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:39.678 15:13:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:39.678 15:13:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:39.678 15:13:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:39.678 15:13:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:39.678 15:13:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.678 15:13:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:39.678 15:13:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:39.678 "name": "Existed_Raid", 00:08:39.678 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:39.678 "strip_size_kb": 64, 00:08:39.678 "state": "configuring", 00:08:39.678 "raid_level": "concat", 00:08:39.678 "superblock": false, 00:08:39.678 "num_base_bdevs": 3, 00:08:39.678 "num_base_bdevs_discovered": 0, 00:08:39.678 "num_base_bdevs_operational": 3, 00:08:39.678 "base_bdevs_list": [ 00:08:39.678 { 00:08:39.678 "name": "BaseBdev1", 00:08:39.678 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:39.678 "is_configured": false, 00:08:39.678 "data_offset": 0, 00:08:39.678 "data_size": 0 00:08:39.678 }, 00:08:39.678 { 00:08:39.678 "name": "BaseBdev2", 00:08:39.678 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:39.678 "is_configured": false, 00:08:39.678 "data_offset": 0, 00:08:39.678 "data_size": 0 00:08:39.678 }, 00:08:39.678 { 00:08:39.678 "name": "BaseBdev3", 00:08:39.678 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:39.678 "is_configured": false, 00:08:39.678 "data_offset": 0, 00:08:39.678 "data_size": 0 00:08:39.678 } 00:08:39.678 ] 00:08:39.678 }' 00:08:39.678 15:13:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:39.678 15:13:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.249 [2024-11-27 15:13:08.115033] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:40.249 [2024-11-27 15:13:08.115114] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.249 [2024-11-27 15:13:08.127046] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:40.249 [2024-11-27 15:13:08.127122] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:40.249 [2024-11-27 15:13:08.127148] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:40.249 [2024-11-27 15:13:08.127170] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:40.249 [2024-11-27 15:13:08.127188] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:40.249 [2024-11-27 15:13:08.127209] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.249 [2024-11-27 15:13:08.148299] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:40.249 BaseBdev1 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.249 [ 00:08:40.249 { 00:08:40.249 "name": "BaseBdev1", 00:08:40.249 "aliases": [ 00:08:40.249 "8d310753-c7a1-4d4d-97e7-778e8b851940" 00:08:40.249 ], 00:08:40.249 "product_name": "Malloc disk", 00:08:40.249 "block_size": 512, 00:08:40.249 "num_blocks": 65536, 00:08:40.249 "uuid": "8d310753-c7a1-4d4d-97e7-778e8b851940", 00:08:40.249 "assigned_rate_limits": { 00:08:40.249 "rw_ios_per_sec": 0, 00:08:40.249 "rw_mbytes_per_sec": 0, 00:08:40.249 "r_mbytes_per_sec": 0, 00:08:40.249 "w_mbytes_per_sec": 0 00:08:40.249 }, 00:08:40.249 "claimed": true, 00:08:40.249 "claim_type": "exclusive_write", 00:08:40.249 "zoned": false, 00:08:40.249 "supported_io_types": { 00:08:40.249 "read": true, 00:08:40.249 "write": true, 00:08:40.249 "unmap": true, 00:08:40.249 "flush": true, 00:08:40.249 "reset": true, 00:08:40.249 "nvme_admin": false, 00:08:40.249 "nvme_io": false, 00:08:40.249 "nvme_io_md": false, 00:08:40.249 "write_zeroes": true, 00:08:40.249 "zcopy": true, 00:08:40.249 "get_zone_info": false, 00:08:40.249 "zone_management": false, 00:08:40.249 "zone_append": false, 00:08:40.249 "compare": false, 00:08:40.249 "compare_and_write": false, 00:08:40.249 "abort": true, 00:08:40.249 "seek_hole": false, 00:08:40.249 "seek_data": false, 00:08:40.249 "copy": true, 00:08:40.249 "nvme_iov_md": false 00:08:40.249 }, 00:08:40.249 "memory_domains": [ 00:08:40.249 { 00:08:40.249 "dma_device_id": "system", 00:08:40.249 "dma_device_type": 1 00:08:40.249 }, 00:08:40.249 { 00:08:40.249 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:40.249 "dma_device_type": 2 00:08:40.249 } 00:08:40.249 ], 00:08:40.249 "driver_specific": {} 00:08:40.249 } 00:08:40.249 ] 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:40.249 "name": "Existed_Raid", 00:08:40.249 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:40.249 "strip_size_kb": 64, 00:08:40.249 "state": "configuring", 00:08:40.249 "raid_level": "concat", 00:08:40.249 "superblock": false, 00:08:40.249 "num_base_bdevs": 3, 00:08:40.249 "num_base_bdevs_discovered": 1, 00:08:40.249 "num_base_bdevs_operational": 3, 00:08:40.249 "base_bdevs_list": [ 00:08:40.249 { 00:08:40.249 "name": "BaseBdev1", 00:08:40.249 "uuid": "8d310753-c7a1-4d4d-97e7-778e8b851940", 00:08:40.249 "is_configured": true, 00:08:40.249 "data_offset": 0, 00:08:40.249 "data_size": 65536 00:08:40.249 }, 00:08:40.249 { 00:08:40.249 "name": "BaseBdev2", 00:08:40.249 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:40.249 "is_configured": false, 00:08:40.249 "data_offset": 0, 00:08:40.249 "data_size": 0 00:08:40.249 }, 00:08:40.249 { 00:08:40.249 "name": "BaseBdev3", 00:08:40.249 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:40.249 "is_configured": false, 00:08:40.249 "data_offset": 0, 00:08:40.249 "data_size": 0 00:08:40.249 } 00:08:40.249 ] 00:08:40.249 }' 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:40.249 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.819 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:40.819 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.819 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.819 [2024-11-27 15:13:08.643504] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:40.819 [2024-11-27 15:13:08.643606] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:08:40.819 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.819 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:40.819 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.819 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.819 [2024-11-27 15:13:08.655501] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:40.819 [2024-11-27 15:13:08.657355] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:40.819 [2024-11-27 15:13:08.657433] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:40.819 [2024-11-27 15:13:08.657460] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:40.819 [2024-11-27 15:13:08.657485] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:40.819 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.819 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:40.819 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:40.819 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:40.819 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:40.819 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:40.819 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:40.819 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:40.819 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:40.819 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:40.819 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:40.819 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:40.819 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:40.819 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:40.819 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.819 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:40.819 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.819 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.819 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:40.819 "name": "Existed_Raid", 00:08:40.819 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:40.819 "strip_size_kb": 64, 00:08:40.819 "state": "configuring", 00:08:40.819 "raid_level": "concat", 00:08:40.819 "superblock": false, 00:08:40.819 "num_base_bdevs": 3, 00:08:40.819 "num_base_bdevs_discovered": 1, 00:08:40.819 "num_base_bdevs_operational": 3, 00:08:40.819 "base_bdevs_list": [ 00:08:40.819 { 00:08:40.819 "name": "BaseBdev1", 00:08:40.819 "uuid": "8d310753-c7a1-4d4d-97e7-778e8b851940", 00:08:40.819 "is_configured": true, 00:08:40.819 "data_offset": 0, 00:08:40.819 "data_size": 65536 00:08:40.819 }, 00:08:40.819 { 00:08:40.819 "name": "BaseBdev2", 00:08:40.819 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:40.819 "is_configured": false, 00:08:40.819 "data_offset": 0, 00:08:40.819 "data_size": 0 00:08:40.819 }, 00:08:40.819 { 00:08:40.819 "name": "BaseBdev3", 00:08:40.819 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:40.819 "is_configured": false, 00:08:40.819 "data_offset": 0, 00:08:40.819 "data_size": 0 00:08:40.819 } 00:08:40.819 ] 00:08:40.819 }' 00:08:40.819 15:13:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:40.819 15:13:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.079 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:41.079 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:41.079 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.079 [2024-11-27 15:13:09.073992] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:41.079 BaseBdev2 00:08:41.079 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:41.079 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:41.079 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:41.079 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:41.079 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:41.079 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:41.079 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:41.079 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:41.079 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:41.079 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.079 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:41.079 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:41.079 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:41.079 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.079 [ 00:08:41.079 { 00:08:41.079 "name": "BaseBdev2", 00:08:41.079 "aliases": [ 00:08:41.079 "8e0928ba-4894-4a57-b776-3d72f26c4624" 00:08:41.079 ], 00:08:41.079 "product_name": "Malloc disk", 00:08:41.079 "block_size": 512, 00:08:41.079 "num_blocks": 65536, 00:08:41.079 "uuid": "8e0928ba-4894-4a57-b776-3d72f26c4624", 00:08:41.079 "assigned_rate_limits": { 00:08:41.079 "rw_ios_per_sec": 0, 00:08:41.079 "rw_mbytes_per_sec": 0, 00:08:41.079 "r_mbytes_per_sec": 0, 00:08:41.079 "w_mbytes_per_sec": 0 00:08:41.079 }, 00:08:41.079 "claimed": true, 00:08:41.079 "claim_type": "exclusive_write", 00:08:41.079 "zoned": false, 00:08:41.079 "supported_io_types": { 00:08:41.079 "read": true, 00:08:41.079 "write": true, 00:08:41.079 "unmap": true, 00:08:41.079 "flush": true, 00:08:41.079 "reset": true, 00:08:41.079 "nvme_admin": false, 00:08:41.079 "nvme_io": false, 00:08:41.079 "nvme_io_md": false, 00:08:41.079 "write_zeroes": true, 00:08:41.079 "zcopy": true, 00:08:41.079 "get_zone_info": false, 00:08:41.079 "zone_management": false, 00:08:41.079 "zone_append": false, 00:08:41.079 "compare": false, 00:08:41.079 "compare_and_write": false, 00:08:41.079 "abort": true, 00:08:41.079 "seek_hole": false, 00:08:41.079 "seek_data": false, 00:08:41.079 "copy": true, 00:08:41.079 "nvme_iov_md": false 00:08:41.079 }, 00:08:41.079 "memory_domains": [ 00:08:41.079 { 00:08:41.079 "dma_device_id": "system", 00:08:41.079 "dma_device_type": 1 00:08:41.079 }, 00:08:41.079 { 00:08:41.079 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:41.079 "dma_device_type": 2 00:08:41.079 } 00:08:41.079 ], 00:08:41.079 "driver_specific": {} 00:08:41.079 } 00:08:41.079 ] 00:08:41.079 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:41.079 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:41.079 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:41.079 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:41.079 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:41.079 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:41.079 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:41.079 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:41.080 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:41.080 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:41.080 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:41.080 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:41.080 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:41.080 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:41.080 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:41.080 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:41.080 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:41.080 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.080 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:41.080 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:41.080 "name": "Existed_Raid", 00:08:41.080 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:41.080 "strip_size_kb": 64, 00:08:41.080 "state": "configuring", 00:08:41.080 "raid_level": "concat", 00:08:41.080 "superblock": false, 00:08:41.080 "num_base_bdevs": 3, 00:08:41.080 "num_base_bdevs_discovered": 2, 00:08:41.080 "num_base_bdevs_operational": 3, 00:08:41.080 "base_bdevs_list": [ 00:08:41.080 { 00:08:41.080 "name": "BaseBdev1", 00:08:41.080 "uuid": "8d310753-c7a1-4d4d-97e7-778e8b851940", 00:08:41.080 "is_configured": true, 00:08:41.080 "data_offset": 0, 00:08:41.080 "data_size": 65536 00:08:41.080 }, 00:08:41.080 { 00:08:41.080 "name": "BaseBdev2", 00:08:41.080 "uuid": "8e0928ba-4894-4a57-b776-3d72f26c4624", 00:08:41.080 "is_configured": true, 00:08:41.080 "data_offset": 0, 00:08:41.080 "data_size": 65536 00:08:41.080 }, 00:08:41.080 { 00:08:41.080 "name": "BaseBdev3", 00:08:41.080 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:41.080 "is_configured": false, 00:08:41.080 "data_offset": 0, 00:08:41.080 "data_size": 0 00:08:41.080 } 00:08:41.080 ] 00:08:41.080 }' 00:08:41.080 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:41.080 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.648 [2024-11-27 15:13:09.548058] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:41.648 [2024-11-27 15:13:09.548181] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:08:41.648 [2024-11-27 15:13:09.548217] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:08:41.648 [2024-11-27 15:13:09.548592] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:08:41.648 [2024-11-27 15:13:09.548820] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:08:41.648 [2024-11-27 15:13:09.548873] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:08:41.648 [2024-11-27 15:13:09.549172] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:41.648 BaseBdev3 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.648 [ 00:08:41.648 { 00:08:41.648 "name": "BaseBdev3", 00:08:41.648 "aliases": [ 00:08:41.648 "9be874da-1b8c-45fc-a4bd-cca08d1a1db3" 00:08:41.648 ], 00:08:41.648 "product_name": "Malloc disk", 00:08:41.648 "block_size": 512, 00:08:41.648 "num_blocks": 65536, 00:08:41.648 "uuid": "9be874da-1b8c-45fc-a4bd-cca08d1a1db3", 00:08:41.648 "assigned_rate_limits": { 00:08:41.648 "rw_ios_per_sec": 0, 00:08:41.648 "rw_mbytes_per_sec": 0, 00:08:41.648 "r_mbytes_per_sec": 0, 00:08:41.648 "w_mbytes_per_sec": 0 00:08:41.648 }, 00:08:41.648 "claimed": true, 00:08:41.648 "claim_type": "exclusive_write", 00:08:41.648 "zoned": false, 00:08:41.648 "supported_io_types": { 00:08:41.648 "read": true, 00:08:41.648 "write": true, 00:08:41.648 "unmap": true, 00:08:41.648 "flush": true, 00:08:41.648 "reset": true, 00:08:41.648 "nvme_admin": false, 00:08:41.648 "nvme_io": false, 00:08:41.648 "nvme_io_md": false, 00:08:41.648 "write_zeroes": true, 00:08:41.648 "zcopy": true, 00:08:41.648 "get_zone_info": false, 00:08:41.648 "zone_management": false, 00:08:41.648 "zone_append": false, 00:08:41.648 "compare": false, 00:08:41.648 "compare_and_write": false, 00:08:41.648 "abort": true, 00:08:41.648 "seek_hole": false, 00:08:41.648 "seek_data": false, 00:08:41.648 "copy": true, 00:08:41.648 "nvme_iov_md": false 00:08:41.648 }, 00:08:41.648 "memory_domains": [ 00:08:41.648 { 00:08:41.648 "dma_device_id": "system", 00:08:41.648 "dma_device_type": 1 00:08:41.648 }, 00:08:41.648 { 00:08:41.648 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:41.648 "dma_device_type": 2 00:08:41.648 } 00:08:41.648 ], 00:08:41.648 "driver_specific": {} 00:08:41.648 } 00:08:41.648 ] 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:41.648 "name": "Existed_Raid", 00:08:41.648 "uuid": "eea64fa4-ef7c-42fb-9a09-1d58562b8777", 00:08:41.648 "strip_size_kb": 64, 00:08:41.648 "state": "online", 00:08:41.648 "raid_level": "concat", 00:08:41.648 "superblock": false, 00:08:41.648 "num_base_bdevs": 3, 00:08:41.648 "num_base_bdevs_discovered": 3, 00:08:41.648 "num_base_bdevs_operational": 3, 00:08:41.648 "base_bdevs_list": [ 00:08:41.648 { 00:08:41.648 "name": "BaseBdev1", 00:08:41.648 "uuid": "8d310753-c7a1-4d4d-97e7-778e8b851940", 00:08:41.648 "is_configured": true, 00:08:41.648 "data_offset": 0, 00:08:41.648 "data_size": 65536 00:08:41.648 }, 00:08:41.648 { 00:08:41.648 "name": "BaseBdev2", 00:08:41.648 "uuid": "8e0928ba-4894-4a57-b776-3d72f26c4624", 00:08:41.648 "is_configured": true, 00:08:41.648 "data_offset": 0, 00:08:41.648 "data_size": 65536 00:08:41.648 }, 00:08:41.648 { 00:08:41.648 "name": "BaseBdev3", 00:08:41.648 "uuid": "9be874da-1b8c-45fc-a4bd-cca08d1a1db3", 00:08:41.648 "is_configured": true, 00:08:41.648 "data_offset": 0, 00:08:41.648 "data_size": 65536 00:08:41.648 } 00:08:41.648 ] 00:08:41.648 }' 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:41.648 15:13:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.218 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:42.218 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:42.218 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:42.218 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:42.218 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:42.218 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:42.218 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:42.218 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:42.218 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.218 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.218 [2024-11-27 15:13:10.055822] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:42.218 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.218 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:42.218 "name": "Existed_Raid", 00:08:42.218 "aliases": [ 00:08:42.218 "eea64fa4-ef7c-42fb-9a09-1d58562b8777" 00:08:42.218 ], 00:08:42.218 "product_name": "Raid Volume", 00:08:42.218 "block_size": 512, 00:08:42.218 "num_blocks": 196608, 00:08:42.218 "uuid": "eea64fa4-ef7c-42fb-9a09-1d58562b8777", 00:08:42.218 "assigned_rate_limits": { 00:08:42.218 "rw_ios_per_sec": 0, 00:08:42.218 "rw_mbytes_per_sec": 0, 00:08:42.218 "r_mbytes_per_sec": 0, 00:08:42.218 "w_mbytes_per_sec": 0 00:08:42.218 }, 00:08:42.218 "claimed": false, 00:08:42.218 "zoned": false, 00:08:42.218 "supported_io_types": { 00:08:42.218 "read": true, 00:08:42.218 "write": true, 00:08:42.218 "unmap": true, 00:08:42.218 "flush": true, 00:08:42.218 "reset": true, 00:08:42.218 "nvme_admin": false, 00:08:42.218 "nvme_io": false, 00:08:42.218 "nvme_io_md": false, 00:08:42.218 "write_zeroes": true, 00:08:42.218 "zcopy": false, 00:08:42.218 "get_zone_info": false, 00:08:42.218 "zone_management": false, 00:08:42.218 "zone_append": false, 00:08:42.218 "compare": false, 00:08:42.218 "compare_and_write": false, 00:08:42.218 "abort": false, 00:08:42.218 "seek_hole": false, 00:08:42.218 "seek_data": false, 00:08:42.218 "copy": false, 00:08:42.218 "nvme_iov_md": false 00:08:42.218 }, 00:08:42.218 "memory_domains": [ 00:08:42.218 { 00:08:42.218 "dma_device_id": "system", 00:08:42.218 "dma_device_type": 1 00:08:42.218 }, 00:08:42.218 { 00:08:42.218 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:42.218 "dma_device_type": 2 00:08:42.218 }, 00:08:42.218 { 00:08:42.218 "dma_device_id": "system", 00:08:42.218 "dma_device_type": 1 00:08:42.218 }, 00:08:42.218 { 00:08:42.218 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:42.218 "dma_device_type": 2 00:08:42.218 }, 00:08:42.218 { 00:08:42.218 "dma_device_id": "system", 00:08:42.218 "dma_device_type": 1 00:08:42.218 }, 00:08:42.218 { 00:08:42.218 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:42.218 "dma_device_type": 2 00:08:42.218 } 00:08:42.218 ], 00:08:42.218 "driver_specific": { 00:08:42.218 "raid": { 00:08:42.218 "uuid": "eea64fa4-ef7c-42fb-9a09-1d58562b8777", 00:08:42.218 "strip_size_kb": 64, 00:08:42.218 "state": "online", 00:08:42.218 "raid_level": "concat", 00:08:42.218 "superblock": false, 00:08:42.218 "num_base_bdevs": 3, 00:08:42.218 "num_base_bdevs_discovered": 3, 00:08:42.218 "num_base_bdevs_operational": 3, 00:08:42.218 "base_bdevs_list": [ 00:08:42.218 { 00:08:42.218 "name": "BaseBdev1", 00:08:42.218 "uuid": "8d310753-c7a1-4d4d-97e7-778e8b851940", 00:08:42.218 "is_configured": true, 00:08:42.218 "data_offset": 0, 00:08:42.218 "data_size": 65536 00:08:42.218 }, 00:08:42.218 { 00:08:42.218 "name": "BaseBdev2", 00:08:42.218 "uuid": "8e0928ba-4894-4a57-b776-3d72f26c4624", 00:08:42.218 "is_configured": true, 00:08:42.218 "data_offset": 0, 00:08:42.218 "data_size": 65536 00:08:42.218 }, 00:08:42.218 { 00:08:42.218 "name": "BaseBdev3", 00:08:42.218 "uuid": "9be874da-1b8c-45fc-a4bd-cca08d1a1db3", 00:08:42.218 "is_configured": true, 00:08:42.218 "data_offset": 0, 00:08:42.218 "data_size": 65536 00:08:42.218 } 00:08:42.218 ] 00:08:42.218 } 00:08:42.218 } 00:08:42.218 }' 00:08:42.218 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:42.218 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:42.218 BaseBdev2 00:08:42.218 BaseBdev3' 00:08:42.218 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:42.218 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:42.218 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:42.218 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:42.218 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:42.218 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.218 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.219 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.219 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:42.219 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:42.219 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:42.219 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:42.219 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.219 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:42.219 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.219 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.219 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:42.219 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:42.219 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:42.219 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:42.219 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:42.219 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.219 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.219 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.219 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:42.219 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:42.219 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:42.219 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.219 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.479 [2024-11-27 15:13:10.323069] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:42.479 [2024-11-27 15:13:10.323141] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:42.479 [2024-11-27 15:13:10.323208] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:42.479 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.479 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:42.479 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:08:42.479 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:42.479 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:42.479 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:08:42.479 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 2 00:08:42.479 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:42.479 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:08:42.479 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:42.479 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:42.479 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:42.479 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:42.479 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:42.479 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:42.479 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:42.479 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:42.479 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.479 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.479 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:42.479 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.479 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:42.479 "name": "Existed_Raid", 00:08:42.479 "uuid": "eea64fa4-ef7c-42fb-9a09-1d58562b8777", 00:08:42.479 "strip_size_kb": 64, 00:08:42.479 "state": "offline", 00:08:42.479 "raid_level": "concat", 00:08:42.479 "superblock": false, 00:08:42.479 "num_base_bdevs": 3, 00:08:42.479 "num_base_bdevs_discovered": 2, 00:08:42.479 "num_base_bdevs_operational": 2, 00:08:42.479 "base_bdevs_list": [ 00:08:42.479 { 00:08:42.479 "name": null, 00:08:42.479 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:42.479 "is_configured": false, 00:08:42.479 "data_offset": 0, 00:08:42.479 "data_size": 65536 00:08:42.479 }, 00:08:42.479 { 00:08:42.479 "name": "BaseBdev2", 00:08:42.479 "uuid": "8e0928ba-4894-4a57-b776-3d72f26c4624", 00:08:42.479 "is_configured": true, 00:08:42.479 "data_offset": 0, 00:08:42.479 "data_size": 65536 00:08:42.479 }, 00:08:42.479 { 00:08:42.479 "name": "BaseBdev3", 00:08:42.479 "uuid": "9be874da-1b8c-45fc-a4bd-cca08d1a1db3", 00:08:42.479 "is_configured": true, 00:08:42.479 "data_offset": 0, 00:08:42.479 "data_size": 65536 00:08:42.479 } 00:08:42.479 ] 00:08:42.479 }' 00:08:42.479 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:42.479 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.740 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:42.740 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:42.740 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:42.740 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.740 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:42.740 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.740 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.740 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:42.740 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:42.740 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:42.740 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.740 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.740 [2024-11-27 15:13:10.817911] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:42.740 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.740 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:42.740 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:42.740 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:42.740 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:42.740 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.740 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.000 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.000 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:43.000 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:43.000 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:08:43.000 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.000 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.000 [2024-11-27 15:13:10.889085] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:43.000 [2024-11-27 15:13:10.889176] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:08:43.000 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.000 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.001 BaseBdev2 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.001 [ 00:08:43.001 { 00:08:43.001 "name": "BaseBdev2", 00:08:43.001 "aliases": [ 00:08:43.001 "db2b836f-19d8-4677-b66b-d9b2f1fa2390" 00:08:43.001 ], 00:08:43.001 "product_name": "Malloc disk", 00:08:43.001 "block_size": 512, 00:08:43.001 "num_blocks": 65536, 00:08:43.001 "uuid": "db2b836f-19d8-4677-b66b-d9b2f1fa2390", 00:08:43.001 "assigned_rate_limits": { 00:08:43.001 "rw_ios_per_sec": 0, 00:08:43.001 "rw_mbytes_per_sec": 0, 00:08:43.001 "r_mbytes_per_sec": 0, 00:08:43.001 "w_mbytes_per_sec": 0 00:08:43.001 }, 00:08:43.001 "claimed": false, 00:08:43.001 "zoned": false, 00:08:43.001 "supported_io_types": { 00:08:43.001 "read": true, 00:08:43.001 "write": true, 00:08:43.001 "unmap": true, 00:08:43.001 "flush": true, 00:08:43.001 "reset": true, 00:08:43.001 "nvme_admin": false, 00:08:43.001 "nvme_io": false, 00:08:43.001 "nvme_io_md": false, 00:08:43.001 "write_zeroes": true, 00:08:43.001 "zcopy": true, 00:08:43.001 "get_zone_info": false, 00:08:43.001 "zone_management": false, 00:08:43.001 "zone_append": false, 00:08:43.001 "compare": false, 00:08:43.001 "compare_and_write": false, 00:08:43.001 "abort": true, 00:08:43.001 "seek_hole": false, 00:08:43.001 "seek_data": false, 00:08:43.001 "copy": true, 00:08:43.001 "nvme_iov_md": false 00:08:43.001 }, 00:08:43.001 "memory_domains": [ 00:08:43.001 { 00:08:43.001 "dma_device_id": "system", 00:08:43.001 "dma_device_type": 1 00:08:43.001 }, 00:08:43.001 { 00:08:43.001 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:43.001 "dma_device_type": 2 00:08:43.001 } 00:08:43.001 ], 00:08:43.001 "driver_specific": {} 00:08:43.001 } 00:08:43.001 ] 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.001 15:13:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.001 BaseBdev3 00:08:43.001 15:13:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.001 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:08:43.001 15:13:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:08:43.001 15:13:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:43.001 15:13:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:43.001 15:13:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:43.001 15:13:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:43.001 15:13:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:43.001 15:13:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.001 15:13:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.001 15:13:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.001 15:13:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:43.001 15:13:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.001 15:13:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.001 [ 00:08:43.001 { 00:08:43.001 "name": "BaseBdev3", 00:08:43.001 "aliases": [ 00:08:43.001 "e494fdfd-133f-4c67-8dc6-277b00dff766" 00:08:43.001 ], 00:08:43.001 "product_name": "Malloc disk", 00:08:43.001 "block_size": 512, 00:08:43.001 "num_blocks": 65536, 00:08:43.001 "uuid": "e494fdfd-133f-4c67-8dc6-277b00dff766", 00:08:43.001 "assigned_rate_limits": { 00:08:43.001 "rw_ios_per_sec": 0, 00:08:43.001 "rw_mbytes_per_sec": 0, 00:08:43.001 "r_mbytes_per_sec": 0, 00:08:43.001 "w_mbytes_per_sec": 0 00:08:43.001 }, 00:08:43.001 "claimed": false, 00:08:43.001 "zoned": false, 00:08:43.001 "supported_io_types": { 00:08:43.001 "read": true, 00:08:43.001 "write": true, 00:08:43.001 "unmap": true, 00:08:43.001 "flush": true, 00:08:43.001 "reset": true, 00:08:43.001 "nvme_admin": false, 00:08:43.001 "nvme_io": false, 00:08:43.001 "nvme_io_md": false, 00:08:43.001 "write_zeroes": true, 00:08:43.001 "zcopy": true, 00:08:43.001 "get_zone_info": false, 00:08:43.001 "zone_management": false, 00:08:43.001 "zone_append": false, 00:08:43.001 "compare": false, 00:08:43.001 "compare_and_write": false, 00:08:43.001 "abort": true, 00:08:43.001 "seek_hole": false, 00:08:43.001 "seek_data": false, 00:08:43.001 "copy": true, 00:08:43.001 "nvme_iov_md": false 00:08:43.001 }, 00:08:43.001 "memory_domains": [ 00:08:43.001 { 00:08:43.001 "dma_device_id": "system", 00:08:43.001 "dma_device_type": 1 00:08:43.001 }, 00:08:43.001 { 00:08:43.001 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:43.001 "dma_device_type": 2 00:08:43.001 } 00:08:43.001 ], 00:08:43.001 "driver_specific": {} 00:08:43.001 } 00:08:43.001 ] 00:08:43.001 15:13:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.001 15:13:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:43.001 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:43.001 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:43.001 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:43.001 15:13:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.001 15:13:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.001 [2024-11-27 15:13:11.045036] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:43.001 [2024-11-27 15:13:11.045125] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:43.001 [2024-11-27 15:13:11.045165] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:43.001 [2024-11-27 15:13:11.046872] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:43.001 15:13:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.001 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:43.001 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:43.001 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:43.001 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:43.001 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:43.001 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:43.001 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:43.002 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:43.002 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:43.002 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:43.002 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:43.002 15:13:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.002 15:13:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.002 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:43.002 15:13:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.260 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:43.260 "name": "Existed_Raid", 00:08:43.261 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:43.261 "strip_size_kb": 64, 00:08:43.261 "state": "configuring", 00:08:43.261 "raid_level": "concat", 00:08:43.261 "superblock": false, 00:08:43.261 "num_base_bdevs": 3, 00:08:43.261 "num_base_bdevs_discovered": 2, 00:08:43.261 "num_base_bdevs_operational": 3, 00:08:43.261 "base_bdevs_list": [ 00:08:43.261 { 00:08:43.261 "name": "BaseBdev1", 00:08:43.261 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:43.261 "is_configured": false, 00:08:43.261 "data_offset": 0, 00:08:43.261 "data_size": 0 00:08:43.261 }, 00:08:43.261 { 00:08:43.261 "name": "BaseBdev2", 00:08:43.261 "uuid": "db2b836f-19d8-4677-b66b-d9b2f1fa2390", 00:08:43.261 "is_configured": true, 00:08:43.261 "data_offset": 0, 00:08:43.261 "data_size": 65536 00:08:43.261 }, 00:08:43.261 { 00:08:43.261 "name": "BaseBdev3", 00:08:43.261 "uuid": "e494fdfd-133f-4c67-8dc6-277b00dff766", 00:08:43.261 "is_configured": true, 00:08:43.261 "data_offset": 0, 00:08:43.261 "data_size": 65536 00:08:43.261 } 00:08:43.261 ] 00:08:43.261 }' 00:08:43.261 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:43.261 15:13:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.520 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:08:43.520 15:13:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.520 15:13:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.520 [2024-11-27 15:13:11.532219] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:43.520 15:13:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.520 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:43.520 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:43.520 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:43.520 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:43.520 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:43.520 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:43.520 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:43.520 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:43.520 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:43.520 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:43.520 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:43.520 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:43.520 15:13:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.520 15:13:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.520 15:13:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.520 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:43.520 "name": "Existed_Raid", 00:08:43.520 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:43.520 "strip_size_kb": 64, 00:08:43.520 "state": "configuring", 00:08:43.520 "raid_level": "concat", 00:08:43.520 "superblock": false, 00:08:43.520 "num_base_bdevs": 3, 00:08:43.520 "num_base_bdevs_discovered": 1, 00:08:43.520 "num_base_bdevs_operational": 3, 00:08:43.520 "base_bdevs_list": [ 00:08:43.520 { 00:08:43.520 "name": "BaseBdev1", 00:08:43.520 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:43.520 "is_configured": false, 00:08:43.520 "data_offset": 0, 00:08:43.520 "data_size": 0 00:08:43.520 }, 00:08:43.520 { 00:08:43.520 "name": null, 00:08:43.520 "uuid": "db2b836f-19d8-4677-b66b-d9b2f1fa2390", 00:08:43.520 "is_configured": false, 00:08:43.520 "data_offset": 0, 00:08:43.520 "data_size": 65536 00:08:43.520 }, 00:08:43.520 { 00:08:43.520 "name": "BaseBdev3", 00:08:43.520 "uuid": "e494fdfd-133f-4c67-8dc6-277b00dff766", 00:08:43.520 "is_configured": true, 00:08:43.520 "data_offset": 0, 00:08:43.520 "data_size": 65536 00:08:43.520 } 00:08:43.520 ] 00:08:43.520 }' 00:08:43.521 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:43.521 15:13:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.090 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:44.090 15:13:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:44.090 15:13:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.090 15:13:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.090 15:13:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.090 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:08:44.090 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:44.090 15:13:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.090 15:13:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.090 [2024-11-27 15:13:12.018533] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:44.090 BaseBdev1 00:08:44.090 15:13:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.090 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:08:44.090 15:13:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:44.090 15:13:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:44.090 15:13:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:44.090 15:13:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:44.090 15:13:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:44.090 15:13:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:44.090 15:13:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.090 15:13:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.090 15:13:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.090 15:13:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:44.090 15:13:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.090 15:13:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.090 [ 00:08:44.090 { 00:08:44.090 "name": "BaseBdev1", 00:08:44.090 "aliases": [ 00:08:44.090 "60e080f1-c5c8-4fa2-b761-b1d87dee8581" 00:08:44.090 ], 00:08:44.090 "product_name": "Malloc disk", 00:08:44.090 "block_size": 512, 00:08:44.090 "num_blocks": 65536, 00:08:44.090 "uuid": "60e080f1-c5c8-4fa2-b761-b1d87dee8581", 00:08:44.090 "assigned_rate_limits": { 00:08:44.090 "rw_ios_per_sec": 0, 00:08:44.090 "rw_mbytes_per_sec": 0, 00:08:44.090 "r_mbytes_per_sec": 0, 00:08:44.090 "w_mbytes_per_sec": 0 00:08:44.090 }, 00:08:44.090 "claimed": true, 00:08:44.090 "claim_type": "exclusive_write", 00:08:44.090 "zoned": false, 00:08:44.090 "supported_io_types": { 00:08:44.090 "read": true, 00:08:44.090 "write": true, 00:08:44.091 "unmap": true, 00:08:44.091 "flush": true, 00:08:44.091 "reset": true, 00:08:44.091 "nvme_admin": false, 00:08:44.091 "nvme_io": false, 00:08:44.091 "nvme_io_md": false, 00:08:44.091 "write_zeroes": true, 00:08:44.091 "zcopy": true, 00:08:44.091 "get_zone_info": false, 00:08:44.091 "zone_management": false, 00:08:44.091 "zone_append": false, 00:08:44.091 "compare": false, 00:08:44.091 "compare_and_write": false, 00:08:44.091 "abort": true, 00:08:44.091 "seek_hole": false, 00:08:44.091 "seek_data": false, 00:08:44.091 "copy": true, 00:08:44.091 "nvme_iov_md": false 00:08:44.091 }, 00:08:44.091 "memory_domains": [ 00:08:44.091 { 00:08:44.091 "dma_device_id": "system", 00:08:44.091 "dma_device_type": 1 00:08:44.091 }, 00:08:44.091 { 00:08:44.091 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:44.091 "dma_device_type": 2 00:08:44.091 } 00:08:44.091 ], 00:08:44.091 "driver_specific": {} 00:08:44.091 } 00:08:44.091 ] 00:08:44.091 15:13:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.091 15:13:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:44.091 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:44.091 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:44.091 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:44.091 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:44.091 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:44.091 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:44.091 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:44.091 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:44.091 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:44.091 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:44.091 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:44.091 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:44.091 15:13:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.091 15:13:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.091 15:13:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.091 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:44.091 "name": "Existed_Raid", 00:08:44.091 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:44.091 "strip_size_kb": 64, 00:08:44.091 "state": "configuring", 00:08:44.091 "raid_level": "concat", 00:08:44.091 "superblock": false, 00:08:44.091 "num_base_bdevs": 3, 00:08:44.091 "num_base_bdevs_discovered": 2, 00:08:44.091 "num_base_bdevs_operational": 3, 00:08:44.091 "base_bdevs_list": [ 00:08:44.091 { 00:08:44.091 "name": "BaseBdev1", 00:08:44.091 "uuid": "60e080f1-c5c8-4fa2-b761-b1d87dee8581", 00:08:44.091 "is_configured": true, 00:08:44.091 "data_offset": 0, 00:08:44.091 "data_size": 65536 00:08:44.091 }, 00:08:44.091 { 00:08:44.091 "name": null, 00:08:44.091 "uuid": "db2b836f-19d8-4677-b66b-d9b2f1fa2390", 00:08:44.091 "is_configured": false, 00:08:44.091 "data_offset": 0, 00:08:44.091 "data_size": 65536 00:08:44.091 }, 00:08:44.091 { 00:08:44.091 "name": "BaseBdev3", 00:08:44.091 "uuid": "e494fdfd-133f-4c67-8dc6-277b00dff766", 00:08:44.091 "is_configured": true, 00:08:44.091 "data_offset": 0, 00:08:44.091 "data_size": 65536 00:08:44.091 } 00:08:44.091 ] 00:08:44.091 }' 00:08:44.091 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:44.091 15:13:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.660 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:44.661 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:44.661 15:13:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.661 15:13:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.661 15:13:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.661 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:08:44.661 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:08:44.661 15:13:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.661 15:13:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.661 [2024-11-27 15:13:12.553678] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:44.661 15:13:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.661 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:44.661 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:44.661 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:44.661 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:44.661 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:44.661 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:44.661 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:44.661 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:44.661 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:44.661 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:44.661 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:44.661 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:44.661 15:13:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.661 15:13:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.661 15:13:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.661 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:44.661 "name": "Existed_Raid", 00:08:44.661 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:44.661 "strip_size_kb": 64, 00:08:44.661 "state": "configuring", 00:08:44.661 "raid_level": "concat", 00:08:44.661 "superblock": false, 00:08:44.661 "num_base_bdevs": 3, 00:08:44.661 "num_base_bdevs_discovered": 1, 00:08:44.661 "num_base_bdevs_operational": 3, 00:08:44.661 "base_bdevs_list": [ 00:08:44.661 { 00:08:44.661 "name": "BaseBdev1", 00:08:44.661 "uuid": "60e080f1-c5c8-4fa2-b761-b1d87dee8581", 00:08:44.661 "is_configured": true, 00:08:44.661 "data_offset": 0, 00:08:44.661 "data_size": 65536 00:08:44.661 }, 00:08:44.661 { 00:08:44.661 "name": null, 00:08:44.661 "uuid": "db2b836f-19d8-4677-b66b-d9b2f1fa2390", 00:08:44.661 "is_configured": false, 00:08:44.661 "data_offset": 0, 00:08:44.661 "data_size": 65536 00:08:44.661 }, 00:08:44.661 { 00:08:44.661 "name": null, 00:08:44.661 "uuid": "e494fdfd-133f-4c67-8dc6-277b00dff766", 00:08:44.661 "is_configured": false, 00:08:44.661 "data_offset": 0, 00:08:44.661 "data_size": 65536 00:08:44.661 } 00:08:44.661 ] 00:08:44.661 }' 00:08:44.661 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:44.661 15:13:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.920 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:44.921 15:13:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:44.921 15:13:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.921 15:13:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.921 15:13:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:45.180 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:08:45.180 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:08:45.180 15:13:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:45.180 15:13:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.180 [2024-11-27 15:13:13.052821] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:45.180 15:13:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:45.180 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:45.180 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:45.180 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:45.180 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:45.180 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:45.180 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:45.180 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:45.180 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:45.180 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:45.180 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:45.180 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:45.180 15:13:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:45.180 15:13:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.180 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:45.180 15:13:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:45.180 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:45.180 "name": "Existed_Raid", 00:08:45.180 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:45.180 "strip_size_kb": 64, 00:08:45.180 "state": "configuring", 00:08:45.180 "raid_level": "concat", 00:08:45.180 "superblock": false, 00:08:45.180 "num_base_bdevs": 3, 00:08:45.180 "num_base_bdevs_discovered": 2, 00:08:45.180 "num_base_bdevs_operational": 3, 00:08:45.180 "base_bdevs_list": [ 00:08:45.180 { 00:08:45.180 "name": "BaseBdev1", 00:08:45.180 "uuid": "60e080f1-c5c8-4fa2-b761-b1d87dee8581", 00:08:45.180 "is_configured": true, 00:08:45.180 "data_offset": 0, 00:08:45.180 "data_size": 65536 00:08:45.180 }, 00:08:45.180 { 00:08:45.180 "name": null, 00:08:45.180 "uuid": "db2b836f-19d8-4677-b66b-d9b2f1fa2390", 00:08:45.180 "is_configured": false, 00:08:45.180 "data_offset": 0, 00:08:45.180 "data_size": 65536 00:08:45.180 }, 00:08:45.180 { 00:08:45.180 "name": "BaseBdev3", 00:08:45.180 "uuid": "e494fdfd-133f-4c67-8dc6-277b00dff766", 00:08:45.180 "is_configured": true, 00:08:45.180 "data_offset": 0, 00:08:45.180 "data_size": 65536 00:08:45.180 } 00:08:45.180 ] 00:08:45.180 }' 00:08:45.180 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:45.180 15:13:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.439 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:45.439 15:13:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:45.439 15:13:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.439 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:45.439 15:13:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:45.439 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:08:45.439 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:45.439 15:13:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:45.439 15:13:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.439 [2024-11-27 15:13:13.540054] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:45.700 15:13:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:45.700 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:45.700 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:45.700 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:45.700 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:45.700 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:45.700 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:45.700 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:45.700 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:45.700 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:45.700 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:45.700 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:45.700 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:45.700 15:13:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:45.700 15:13:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.700 15:13:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:45.700 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:45.700 "name": "Existed_Raid", 00:08:45.700 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:45.700 "strip_size_kb": 64, 00:08:45.700 "state": "configuring", 00:08:45.700 "raid_level": "concat", 00:08:45.700 "superblock": false, 00:08:45.700 "num_base_bdevs": 3, 00:08:45.700 "num_base_bdevs_discovered": 1, 00:08:45.700 "num_base_bdevs_operational": 3, 00:08:45.700 "base_bdevs_list": [ 00:08:45.700 { 00:08:45.700 "name": null, 00:08:45.700 "uuid": "60e080f1-c5c8-4fa2-b761-b1d87dee8581", 00:08:45.700 "is_configured": false, 00:08:45.700 "data_offset": 0, 00:08:45.700 "data_size": 65536 00:08:45.700 }, 00:08:45.700 { 00:08:45.700 "name": null, 00:08:45.700 "uuid": "db2b836f-19d8-4677-b66b-d9b2f1fa2390", 00:08:45.700 "is_configured": false, 00:08:45.700 "data_offset": 0, 00:08:45.700 "data_size": 65536 00:08:45.700 }, 00:08:45.700 { 00:08:45.700 "name": "BaseBdev3", 00:08:45.700 "uuid": "e494fdfd-133f-4c67-8dc6-277b00dff766", 00:08:45.700 "is_configured": true, 00:08:45.700 "data_offset": 0, 00:08:45.700 "data_size": 65536 00:08:45.700 } 00:08:45.700 ] 00:08:45.700 }' 00:08:45.700 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:45.700 15:13:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.960 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:45.960 15:13:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:45.960 15:13:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:45.960 15:13:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.960 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:45.960 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:08:45.960 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:08:45.960 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:45.960 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.960 [2024-11-27 15:13:14.033883] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:45.960 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:45.960 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:45.960 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:45.960 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:45.960 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:45.960 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:45.960 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:45.960 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:45.960 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:45.960 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:45.960 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:45.960 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:45.960 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:45.960 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:45.960 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.225 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.225 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:46.225 "name": "Existed_Raid", 00:08:46.225 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:46.225 "strip_size_kb": 64, 00:08:46.225 "state": "configuring", 00:08:46.225 "raid_level": "concat", 00:08:46.225 "superblock": false, 00:08:46.225 "num_base_bdevs": 3, 00:08:46.225 "num_base_bdevs_discovered": 2, 00:08:46.225 "num_base_bdevs_operational": 3, 00:08:46.225 "base_bdevs_list": [ 00:08:46.225 { 00:08:46.225 "name": null, 00:08:46.225 "uuid": "60e080f1-c5c8-4fa2-b761-b1d87dee8581", 00:08:46.225 "is_configured": false, 00:08:46.225 "data_offset": 0, 00:08:46.225 "data_size": 65536 00:08:46.225 }, 00:08:46.225 { 00:08:46.225 "name": "BaseBdev2", 00:08:46.225 "uuid": "db2b836f-19d8-4677-b66b-d9b2f1fa2390", 00:08:46.225 "is_configured": true, 00:08:46.225 "data_offset": 0, 00:08:46.225 "data_size": 65536 00:08:46.225 }, 00:08:46.225 { 00:08:46.225 "name": "BaseBdev3", 00:08:46.225 "uuid": "e494fdfd-133f-4c67-8dc6-277b00dff766", 00:08:46.225 "is_configured": true, 00:08:46.225 "data_offset": 0, 00:08:46.225 "data_size": 65536 00:08:46.225 } 00:08:46.225 ] 00:08:46.225 }' 00:08:46.225 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:46.225 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.496 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:46.496 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.496 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.496 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:46.496 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.496 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:08:46.496 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:46.496 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:08:46.496 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.496 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.496 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.756 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 60e080f1-c5c8-4fa2-b761-b1d87dee8581 00:08:46.756 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.756 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.756 [2024-11-27 15:13:14.616133] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:08:46.756 [2024-11-27 15:13:14.616233] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:08:46.756 [2024-11-27 15:13:14.616261] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:08:46.756 [2024-11-27 15:13:14.616529] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:08:46.756 [2024-11-27 15:13:14.616691] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:08:46.756 [2024-11-27 15:13:14.616732] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006d00 00:08:46.756 [2024-11-27 15:13:14.616962] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:46.756 NewBaseBdev 00:08:46.756 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.756 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:08:46.756 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:08:46.756 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:46.756 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:46.756 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:46.756 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:46.756 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:46.756 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.756 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.756 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.756 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:08:46.756 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.756 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.756 [ 00:08:46.756 { 00:08:46.756 "name": "NewBaseBdev", 00:08:46.756 "aliases": [ 00:08:46.756 "60e080f1-c5c8-4fa2-b761-b1d87dee8581" 00:08:46.756 ], 00:08:46.756 "product_name": "Malloc disk", 00:08:46.756 "block_size": 512, 00:08:46.756 "num_blocks": 65536, 00:08:46.756 "uuid": "60e080f1-c5c8-4fa2-b761-b1d87dee8581", 00:08:46.756 "assigned_rate_limits": { 00:08:46.756 "rw_ios_per_sec": 0, 00:08:46.756 "rw_mbytes_per_sec": 0, 00:08:46.756 "r_mbytes_per_sec": 0, 00:08:46.756 "w_mbytes_per_sec": 0 00:08:46.756 }, 00:08:46.756 "claimed": true, 00:08:46.756 "claim_type": "exclusive_write", 00:08:46.756 "zoned": false, 00:08:46.756 "supported_io_types": { 00:08:46.756 "read": true, 00:08:46.756 "write": true, 00:08:46.756 "unmap": true, 00:08:46.756 "flush": true, 00:08:46.756 "reset": true, 00:08:46.756 "nvme_admin": false, 00:08:46.756 "nvme_io": false, 00:08:46.756 "nvme_io_md": false, 00:08:46.756 "write_zeroes": true, 00:08:46.756 "zcopy": true, 00:08:46.756 "get_zone_info": false, 00:08:46.756 "zone_management": false, 00:08:46.756 "zone_append": false, 00:08:46.756 "compare": false, 00:08:46.756 "compare_and_write": false, 00:08:46.756 "abort": true, 00:08:46.756 "seek_hole": false, 00:08:46.756 "seek_data": false, 00:08:46.756 "copy": true, 00:08:46.756 "nvme_iov_md": false 00:08:46.756 }, 00:08:46.756 "memory_domains": [ 00:08:46.756 { 00:08:46.756 "dma_device_id": "system", 00:08:46.756 "dma_device_type": 1 00:08:46.756 }, 00:08:46.757 { 00:08:46.757 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:46.757 "dma_device_type": 2 00:08:46.757 } 00:08:46.757 ], 00:08:46.757 "driver_specific": {} 00:08:46.757 } 00:08:46.757 ] 00:08:46.757 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.757 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:46.757 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:08:46.757 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:46.757 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:46.757 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:46.757 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:46.757 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:46.757 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:46.757 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:46.757 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:46.757 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:46.757 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:46.757 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:46.757 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.757 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.757 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.757 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:46.757 "name": "Existed_Raid", 00:08:46.757 "uuid": "31381d01-1e53-45ad-9013-307323922191", 00:08:46.757 "strip_size_kb": 64, 00:08:46.757 "state": "online", 00:08:46.757 "raid_level": "concat", 00:08:46.757 "superblock": false, 00:08:46.757 "num_base_bdevs": 3, 00:08:46.757 "num_base_bdevs_discovered": 3, 00:08:46.757 "num_base_bdevs_operational": 3, 00:08:46.757 "base_bdevs_list": [ 00:08:46.757 { 00:08:46.757 "name": "NewBaseBdev", 00:08:46.757 "uuid": "60e080f1-c5c8-4fa2-b761-b1d87dee8581", 00:08:46.757 "is_configured": true, 00:08:46.757 "data_offset": 0, 00:08:46.757 "data_size": 65536 00:08:46.757 }, 00:08:46.757 { 00:08:46.757 "name": "BaseBdev2", 00:08:46.757 "uuid": "db2b836f-19d8-4677-b66b-d9b2f1fa2390", 00:08:46.757 "is_configured": true, 00:08:46.757 "data_offset": 0, 00:08:46.757 "data_size": 65536 00:08:46.757 }, 00:08:46.757 { 00:08:46.757 "name": "BaseBdev3", 00:08:46.757 "uuid": "e494fdfd-133f-4c67-8dc6-277b00dff766", 00:08:46.757 "is_configured": true, 00:08:46.757 "data_offset": 0, 00:08:46.757 "data_size": 65536 00:08:46.757 } 00:08:46.757 ] 00:08:46.757 }' 00:08:46.757 15:13:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:46.757 15:13:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.327 15:13:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:08:47.327 15:13:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:47.327 15:13:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:47.327 15:13:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:47.327 15:13:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:47.327 15:13:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:47.327 15:13:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:47.327 15:13:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.327 15:13:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.327 15:13:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:47.327 [2024-11-27 15:13:15.140115] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:47.327 15:13:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.327 15:13:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:47.327 "name": "Existed_Raid", 00:08:47.327 "aliases": [ 00:08:47.327 "31381d01-1e53-45ad-9013-307323922191" 00:08:47.327 ], 00:08:47.327 "product_name": "Raid Volume", 00:08:47.327 "block_size": 512, 00:08:47.327 "num_blocks": 196608, 00:08:47.327 "uuid": "31381d01-1e53-45ad-9013-307323922191", 00:08:47.327 "assigned_rate_limits": { 00:08:47.327 "rw_ios_per_sec": 0, 00:08:47.327 "rw_mbytes_per_sec": 0, 00:08:47.327 "r_mbytes_per_sec": 0, 00:08:47.327 "w_mbytes_per_sec": 0 00:08:47.327 }, 00:08:47.327 "claimed": false, 00:08:47.327 "zoned": false, 00:08:47.327 "supported_io_types": { 00:08:47.327 "read": true, 00:08:47.327 "write": true, 00:08:47.327 "unmap": true, 00:08:47.327 "flush": true, 00:08:47.327 "reset": true, 00:08:47.327 "nvme_admin": false, 00:08:47.327 "nvme_io": false, 00:08:47.327 "nvme_io_md": false, 00:08:47.327 "write_zeroes": true, 00:08:47.327 "zcopy": false, 00:08:47.327 "get_zone_info": false, 00:08:47.327 "zone_management": false, 00:08:47.327 "zone_append": false, 00:08:47.327 "compare": false, 00:08:47.327 "compare_and_write": false, 00:08:47.327 "abort": false, 00:08:47.327 "seek_hole": false, 00:08:47.327 "seek_data": false, 00:08:47.327 "copy": false, 00:08:47.327 "nvme_iov_md": false 00:08:47.327 }, 00:08:47.327 "memory_domains": [ 00:08:47.327 { 00:08:47.327 "dma_device_id": "system", 00:08:47.327 "dma_device_type": 1 00:08:47.327 }, 00:08:47.327 { 00:08:47.327 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:47.327 "dma_device_type": 2 00:08:47.327 }, 00:08:47.327 { 00:08:47.327 "dma_device_id": "system", 00:08:47.327 "dma_device_type": 1 00:08:47.327 }, 00:08:47.327 { 00:08:47.327 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:47.327 "dma_device_type": 2 00:08:47.327 }, 00:08:47.327 { 00:08:47.327 "dma_device_id": "system", 00:08:47.327 "dma_device_type": 1 00:08:47.327 }, 00:08:47.327 { 00:08:47.327 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:47.327 "dma_device_type": 2 00:08:47.327 } 00:08:47.327 ], 00:08:47.327 "driver_specific": { 00:08:47.327 "raid": { 00:08:47.327 "uuid": "31381d01-1e53-45ad-9013-307323922191", 00:08:47.327 "strip_size_kb": 64, 00:08:47.327 "state": "online", 00:08:47.327 "raid_level": "concat", 00:08:47.327 "superblock": false, 00:08:47.327 "num_base_bdevs": 3, 00:08:47.327 "num_base_bdevs_discovered": 3, 00:08:47.327 "num_base_bdevs_operational": 3, 00:08:47.327 "base_bdevs_list": [ 00:08:47.327 { 00:08:47.327 "name": "NewBaseBdev", 00:08:47.327 "uuid": "60e080f1-c5c8-4fa2-b761-b1d87dee8581", 00:08:47.327 "is_configured": true, 00:08:47.327 "data_offset": 0, 00:08:47.327 "data_size": 65536 00:08:47.327 }, 00:08:47.327 { 00:08:47.327 "name": "BaseBdev2", 00:08:47.327 "uuid": "db2b836f-19d8-4677-b66b-d9b2f1fa2390", 00:08:47.327 "is_configured": true, 00:08:47.327 "data_offset": 0, 00:08:47.327 "data_size": 65536 00:08:47.327 }, 00:08:47.327 { 00:08:47.327 "name": "BaseBdev3", 00:08:47.327 "uuid": "e494fdfd-133f-4c67-8dc6-277b00dff766", 00:08:47.327 "is_configured": true, 00:08:47.327 "data_offset": 0, 00:08:47.327 "data_size": 65536 00:08:47.328 } 00:08:47.328 ] 00:08:47.328 } 00:08:47.328 } 00:08:47.328 }' 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:08:47.328 BaseBdev2 00:08:47.328 BaseBdev3' 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.328 [2024-11-27 15:13:15.415797] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:47.328 [2024-11-27 15:13:15.415867] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:47.328 [2024-11-27 15:13:15.415967] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:47.328 [2024-11-27 15:13:15.416045] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:47.328 [2024-11-27 15:13:15.416100] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name Existed_Raid, state offline 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 76907 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 76907 ']' 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 76907 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:08:47.328 15:13:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:47.587 15:13:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 76907 00:08:47.588 15:13:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:47.588 15:13:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:47.588 15:13:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 76907' 00:08:47.588 killing process with pid 76907 00:08:47.588 15:13:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 76907 00:08:47.588 [2024-11-27 15:13:15.466608] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:47.588 15:13:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 76907 00:08:47.588 [2024-11-27 15:13:15.497166] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:08:47.847 ************************************ 00:08:47.847 END TEST raid_state_function_test 00:08:47.847 ************************************ 00:08:47.847 00:08:47.847 real 0m8.973s 00:08:47.847 user 0m15.258s 00:08:47.847 sys 0m1.905s 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.847 15:13:15 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 3 true 00:08:47.847 15:13:15 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:47.847 15:13:15 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:47.847 15:13:15 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:47.847 ************************************ 00:08:47.847 START TEST raid_state_function_test_sb 00:08:47.847 ************************************ 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 3 true 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=77512 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:47.847 Process raid pid: 77512 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 77512' 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 77512 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 77512 ']' 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:47.847 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:47.847 15:13:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:47.847 [2024-11-27 15:13:15.894562] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:08:47.847 [2024-11-27 15:13:15.895209] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:48.106 [2024-11-27 15:13:16.070785] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:48.106 [2024-11-27 15:13:16.096240] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:48.107 [2024-11-27 15:13:16.139694] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:48.107 [2024-11-27 15:13:16.139813] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:48.676 15:13:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:48.676 15:13:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:08:48.676 15:13:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:48.676 15:13:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.676 15:13:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:48.676 [2024-11-27 15:13:16.727317] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:48.676 [2024-11-27 15:13:16.727417] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:48.676 [2024-11-27 15:13:16.727447] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:48.676 [2024-11-27 15:13:16.727470] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:48.676 [2024-11-27 15:13:16.727491] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:48.676 [2024-11-27 15:13:16.727513] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:48.676 15:13:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.676 15:13:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:48.676 15:13:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:48.676 15:13:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:48.676 15:13:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:48.676 15:13:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:48.676 15:13:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:48.676 15:13:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:48.676 15:13:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:48.676 15:13:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:48.676 15:13:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:48.676 15:13:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:48.676 15:13:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:48.676 15:13:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.676 15:13:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:48.676 15:13:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.936 15:13:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:48.936 "name": "Existed_Raid", 00:08:48.936 "uuid": "7d133cb3-7fd8-44fb-bdbc-fc24eff3f0a4", 00:08:48.936 "strip_size_kb": 64, 00:08:48.936 "state": "configuring", 00:08:48.936 "raid_level": "concat", 00:08:48.936 "superblock": true, 00:08:48.936 "num_base_bdevs": 3, 00:08:48.936 "num_base_bdevs_discovered": 0, 00:08:48.936 "num_base_bdevs_operational": 3, 00:08:48.936 "base_bdevs_list": [ 00:08:48.936 { 00:08:48.936 "name": "BaseBdev1", 00:08:48.936 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:48.936 "is_configured": false, 00:08:48.936 "data_offset": 0, 00:08:48.936 "data_size": 0 00:08:48.936 }, 00:08:48.936 { 00:08:48.936 "name": "BaseBdev2", 00:08:48.936 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:48.936 "is_configured": false, 00:08:48.936 "data_offset": 0, 00:08:48.936 "data_size": 0 00:08:48.936 }, 00:08:48.936 { 00:08:48.936 "name": "BaseBdev3", 00:08:48.936 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:48.936 "is_configured": false, 00:08:48.936 "data_offset": 0, 00:08:48.936 "data_size": 0 00:08:48.936 } 00:08:48.936 ] 00:08:48.936 }' 00:08:48.936 15:13:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:48.936 15:13:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:49.196 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:49.196 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.196 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:49.196 [2024-11-27 15:13:17.194442] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:49.196 [2024-11-27 15:13:17.194544] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:08:49.196 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.196 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:49.196 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.196 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:49.196 [2024-11-27 15:13:17.206433] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:49.196 [2024-11-27 15:13:17.206510] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:49.196 [2024-11-27 15:13:17.206538] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:49.196 [2024-11-27 15:13:17.206560] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:49.196 [2024-11-27 15:13:17.206577] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:49.196 [2024-11-27 15:13:17.206598] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:49.196 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.196 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:49.196 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.196 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:49.196 [2024-11-27 15:13:17.227441] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:49.196 BaseBdev1 00:08:49.196 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.196 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:49.196 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:49.196 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:49.196 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:49.196 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:49.196 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:49.196 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:49.196 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.196 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:49.196 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.196 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:49.196 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.196 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:49.196 [ 00:08:49.196 { 00:08:49.196 "name": "BaseBdev1", 00:08:49.196 "aliases": [ 00:08:49.196 "9f8d3326-62ce-4f94-80a7-e9a0c7f31cd9" 00:08:49.197 ], 00:08:49.197 "product_name": "Malloc disk", 00:08:49.197 "block_size": 512, 00:08:49.197 "num_blocks": 65536, 00:08:49.197 "uuid": "9f8d3326-62ce-4f94-80a7-e9a0c7f31cd9", 00:08:49.197 "assigned_rate_limits": { 00:08:49.197 "rw_ios_per_sec": 0, 00:08:49.197 "rw_mbytes_per_sec": 0, 00:08:49.197 "r_mbytes_per_sec": 0, 00:08:49.197 "w_mbytes_per_sec": 0 00:08:49.197 }, 00:08:49.197 "claimed": true, 00:08:49.197 "claim_type": "exclusive_write", 00:08:49.197 "zoned": false, 00:08:49.197 "supported_io_types": { 00:08:49.197 "read": true, 00:08:49.197 "write": true, 00:08:49.197 "unmap": true, 00:08:49.197 "flush": true, 00:08:49.197 "reset": true, 00:08:49.197 "nvme_admin": false, 00:08:49.197 "nvme_io": false, 00:08:49.197 "nvme_io_md": false, 00:08:49.197 "write_zeroes": true, 00:08:49.197 "zcopy": true, 00:08:49.197 "get_zone_info": false, 00:08:49.197 "zone_management": false, 00:08:49.197 "zone_append": false, 00:08:49.197 "compare": false, 00:08:49.197 "compare_and_write": false, 00:08:49.197 "abort": true, 00:08:49.197 "seek_hole": false, 00:08:49.197 "seek_data": false, 00:08:49.197 "copy": true, 00:08:49.197 "nvme_iov_md": false 00:08:49.197 }, 00:08:49.197 "memory_domains": [ 00:08:49.197 { 00:08:49.197 "dma_device_id": "system", 00:08:49.197 "dma_device_type": 1 00:08:49.197 }, 00:08:49.197 { 00:08:49.197 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:49.197 "dma_device_type": 2 00:08:49.197 } 00:08:49.197 ], 00:08:49.197 "driver_specific": {} 00:08:49.197 } 00:08:49.197 ] 00:08:49.197 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.197 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:49.197 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:49.197 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:49.197 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:49.197 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:49.197 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:49.197 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:49.197 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:49.197 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:49.197 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:49.197 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:49.197 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:49.197 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:49.197 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.197 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:49.197 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.457 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:49.457 "name": "Existed_Raid", 00:08:49.457 "uuid": "568763b6-b71b-474e-a426-6a2ffcfc90d6", 00:08:49.457 "strip_size_kb": 64, 00:08:49.457 "state": "configuring", 00:08:49.457 "raid_level": "concat", 00:08:49.457 "superblock": true, 00:08:49.457 "num_base_bdevs": 3, 00:08:49.457 "num_base_bdevs_discovered": 1, 00:08:49.457 "num_base_bdevs_operational": 3, 00:08:49.457 "base_bdevs_list": [ 00:08:49.457 { 00:08:49.457 "name": "BaseBdev1", 00:08:49.457 "uuid": "9f8d3326-62ce-4f94-80a7-e9a0c7f31cd9", 00:08:49.457 "is_configured": true, 00:08:49.457 "data_offset": 2048, 00:08:49.457 "data_size": 63488 00:08:49.457 }, 00:08:49.457 { 00:08:49.457 "name": "BaseBdev2", 00:08:49.457 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:49.457 "is_configured": false, 00:08:49.457 "data_offset": 0, 00:08:49.457 "data_size": 0 00:08:49.457 }, 00:08:49.457 { 00:08:49.457 "name": "BaseBdev3", 00:08:49.457 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:49.457 "is_configured": false, 00:08:49.457 "data_offset": 0, 00:08:49.457 "data_size": 0 00:08:49.457 } 00:08:49.457 ] 00:08:49.457 }' 00:08:49.457 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:49.457 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:49.716 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:49.716 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.716 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:49.716 [2024-11-27 15:13:17.750588] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:49.716 [2024-11-27 15:13:17.750687] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:08:49.716 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.716 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:49.716 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.716 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:49.716 [2024-11-27 15:13:17.762596] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:49.716 [2024-11-27 15:13:17.764504] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:49.716 [2024-11-27 15:13:17.764583] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:49.716 [2024-11-27 15:13:17.764613] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:49.716 [2024-11-27 15:13:17.764638] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:49.716 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.716 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:49.716 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:49.716 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:49.716 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:49.716 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:49.716 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:49.716 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:49.716 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:49.716 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:49.716 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:49.716 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:49.716 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:49.716 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:49.716 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.716 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:49.716 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:49.716 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.976 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:49.976 "name": "Existed_Raid", 00:08:49.976 "uuid": "070d3f19-b3f6-4750-bad0-4d82a0ee4ea4", 00:08:49.976 "strip_size_kb": 64, 00:08:49.976 "state": "configuring", 00:08:49.976 "raid_level": "concat", 00:08:49.976 "superblock": true, 00:08:49.976 "num_base_bdevs": 3, 00:08:49.976 "num_base_bdevs_discovered": 1, 00:08:49.976 "num_base_bdevs_operational": 3, 00:08:49.976 "base_bdevs_list": [ 00:08:49.976 { 00:08:49.976 "name": "BaseBdev1", 00:08:49.976 "uuid": "9f8d3326-62ce-4f94-80a7-e9a0c7f31cd9", 00:08:49.976 "is_configured": true, 00:08:49.976 "data_offset": 2048, 00:08:49.976 "data_size": 63488 00:08:49.976 }, 00:08:49.976 { 00:08:49.976 "name": "BaseBdev2", 00:08:49.976 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:49.976 "is_configured": false, 00:08:49.976 "data_offset": 0, 00:08:49.976 "data_size": 0 00:08:49.976 }, 00:08:49.976 { 00:08:49.976 "name": "BaseBdev3", 00:08:49.976 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:49.976 "is_configured": false, 00:08:49.976 "data_offset": 0, 00:08:49.976 "data_size": 0 00:08:49.976 } 00:08:49.976 ] 00:08:49.976 }' 00:08:49.976 15:13:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:49.976 15:13:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:50.235 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:50.235 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.235 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:50.235 [2024-11-27 15:13:18.169100] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:50.235 BaseBdev2 00:08:50.235 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.235 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:50.235 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:50.235 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:50.235 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:50.235 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:50.235 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:50.235 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:50.235 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.235 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:50.235 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.235 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:50.235 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.235 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:50.235 [ 00:08:50.235 { 00:08:50.235 "name": "BaseBdev2", 00:08:50.235 "aliases": [ 00:08:50.235 "43fddafa-ce11-4408-8b6e-1b19dfe9f9f3" 00:08:50.235 ], 00:08:50.235 "product_name": "Malloc disk", 00:08:50.236 "block_size": 512, 00:08:50.236 "num_blocks": 65536, 00:08:50.236 "uuid": "43fddafa-ce11-4408-8b6e-1b19dfe9f9f3", 00:08:50.236 "assigned_rate_limits": { 00:08:50.236 "rw_ios_per_sec": 0, 00:08:50.236 "rw_mbytes_per_sec": 0, 00:08:50.236 "r_mbytes_per_sec": 0, 00:08:50.236 "w_mbytes_per_sec": 0 00:08:50.236 }, 00:08:50.236 "claimed": true, 00:08:50.236 "claim_type": "exclusive_write", 00:08:50.236 "zoned": false, 00:08:50.236 "supported_io_types": { 00:08:50.236 "read": true, 00:08:50.236 "write": true, 00:08:50.236 "unmap": true, 00:08:50.236 "flush": true, 00:08:50.236 "reset": true, 00:08:50.236 "nvme_admin": false, 00:08:50.236 "nvme_io": false, 00:08:50.236 "nvme_io_md": false, 00:08:50.236 "write_zeroes": true, 00:08:50.236 "zcopy": true, 00:08:50.236 "get_zone_info": false, 00:08:50.236 "zone_management": false, 00:08:50.236 "zone_append": false, 00:08:50.236 "compare": false, 00:08:50.236 "compare_and_write": false, 00:08:50.236 "abort": true, 00:08:50.236 "seek_hole": false, 00:08:50.236 "seek_data": false, 00:08:50.236 "copy": true, 00:08:50.236 "nvme_iov_md": false 00:08:50.236 }, 00:08:50.236 "memory_domains": [ 00:08:50.236 { 00:08:50.236 "dma_device_id": "system", 00:08:50.236 "dma_device_type": 1 00:08:50.236 }, 00:08:50.236 { 00:08:50.236 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:50.236 "dma_device_type": 2 00:08:50.236 } 00:08:50.236 ], 00:08:50.236 "driver_specific": {} 00:08:50.236 } 00:08:50.236 ] 00:08:50.236 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.236 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:50.236 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:50.236 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:50.236 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:50.236 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:50.236 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:50.236 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:50.236 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:50.236 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:50.236 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:50.236 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:50.236 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:50.236 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:50.236 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:50.236 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.236 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:50.236 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:50.236 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.236 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:50.236 "name": "Existed_Raid", 00:08:50.236 "uuid": "070d3f19-b3f6-4750-bad0-4d82a0ee4ea4", 00:08:50.236 "strip_size_kb": 64, 00:08:50.236 "state": "configuring", 00:08:50.236 "raid_level": "concat", 00:08:50.236 "superblock": true, 00:08:50.236 "num_base_bdevs": 3, 00:08:50.236 "num_base_bdevs_discovered": 2, 00:08:50.236 "num_base_bdevs_operational": 3, 00:08:50.236 "base_bdevs_list": [ 00:08:50.236 { 00:08:50.236 "name": "BaseBdev1", 00:08:50.236 "uuid": "9f8d3326-62ce-4f94-80a7-e9a0c7f31cd9", 00:08:50.236 "is_configured": true, 00:08:50.236 "data_offset": 2048, 00:08:50.236 "data_size": 63488 00:08:50.236 }, 00:08:50.236 { 00:08:50.236 "name": "BaseBdev2", 00:08:50.236 "uuid": "43fddafa-ce11-4408-8b6e-1b19dfe9f9f3", 00:08:50.236 "is_configured": true, 00:08:50.236 "data_offset": 2048, 00:08:50.236 "data_size": 63488 00:08:50.236 }, 00:08:50.236 { 00:08:50.236 "name": "BaseBdev3", 00:08:50.236 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:50.236 "is_configured": false, 00:08:50.236 "data_offset": 0, 00:08:50.236 "data_size": 0 00:08:50.236 } 00:08:50.236 ] 00:08:50.236 }' 00:08:50.236 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:50.236 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:50.805 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:50.805 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.805 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:50.805 [2024-11-27 15:13:18.721498] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:50.805 [2024-11-27 15:13:18.721716] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:08:50.805 [2024-11-27 15:13:18.721747] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:50.805 [2024-11-27 15:13:18.722113] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:08:50.805 BaseBdev3 00:08:50.805 [2024-11-27 15:13:18.722306] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:08:50.805 [2024-11-27 15:13:18.722327] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:08:50.805 [2024-11-27 15:13:18.722466] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:50.805 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.805 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:08:50.805 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:08:50.805 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:50.805 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:50.805 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:50.805 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:50.805 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:50.805 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.805 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:50.805 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.805 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:50.805 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.805 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:50.805 [ 00:08:50.805 { 00:08:50.805 "name": "BaseBdev3", 00:08:50.805 "aliases": [ 00:08:50.805 "07cc63b6-e3c4-4af3-acd3-84a4e8af7293" 00:08:50.805 ], 00:08:50.805 "product_name": "Malloc disk", 00:08:50.805 "block_size": 512, 00:08:50.805 "num_blocks": 65536, 00:08:50.805 "uuid": "07cc63b6-e3c4-4af3-acd3-84a4e8af7293", 00:08:50.805 "assigned_rate_limits": { 00:08:50.805 "rw_ios_per_sec": 0, 00:08:50.805 "rw_mbytes_per_sec": 0, 00:08:50.805 "r_mbytes_per_sec": 0, 00:08:50.805 "w_mbytes_per_sec": 0 00:08:50.805 }, 00:08:50.805 "claimed": true, 00:08:50.805 "claim_type": "exclusive_write", 00:08:50.805 "zoned": false, 00:08:50.805 "supported_io_types": { 00:08:50.805 "read": true, 00:08:50.805 "write": true, 00:08:50.805 "unmap": true, 00:08:50.805 "flush": true, 00:08:50.805 "reset": true, 00:08:50.805 "nvme_admin": false, 00:08:50.805 "nvme_io": false, 00:08:50.805 "nvme_io_md": false, 00:08:50.805 "write_zeroes": true, 00:08:50.805 "zcopy": true, 00:08:50.805 "get_zone_info": false, 00:08:50.805 "zone_management": false, 00:08:50.806 "zone_append": false, 00:08:50.806 "compare": false, 00:08:50.806 "compare_and_write": false, 00:08:50.806 "abort": true, 00:08:50.806 "seek_hole": false, 00:08:50.806 "seek_data": false, 00:08:50.806 "copy": true, 00:08:50.806 "nvme_iov_md": false 00:08:50.806 }, 00:08:50.806 "memory_domains": [ 00:08:50.806 { 00:08:50.806 "dma_device_id": "system", 00:08:50.806 "dma_device_type": 1 00:08:50.806 }, 00:08:50.806 { 00:08:50.806 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:50.806 "dma_device_type": 2 00:08:50.806 } 00:08:50.806 ], 00:08:50.806 "driver_specific": {} 00:08:50.806 } 00:08:50.806 ] 00:08:50.806 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.806 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:50.806 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:50.806 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:50.806 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:08:50.806 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:50.806 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:50.806 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:50.806 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:50.806 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:50.806 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:50.806 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:50.806 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:50.806 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:50.806 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:50.806 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:50.806 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.806 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:50.806 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.806 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:50.806 "name": "Existed_Raid", 00:08:50.806 "uuid": "070d3f19-b3f6-4750-bad0-4d82a0ee4ea4", 00:08:50.806 "strip_size_kb": 64, 00:08:50.806 "state": "online", 00:08:50.806 "raid_level": "concat", 00:08:50.806 "superblock": true, 00:08:50.806 "num_base_bdevs": 3, 00:08:50.806 "num_base_bdevs_discovered": 3, 00:08:50.806 "num_base_bdevs_operational": 3, 00:08:50.806 "base_bdevs_list": [ 00:08:50.806 { 00:08:50.806 "name": "BaseBdev1", 00:08:50.806 "uuid": "9f8d3326-62ce-4f94-80a7-e9a0c7f31cd9", 00:08:50.806 "is_configured": true, 00:08:50.806 "data_offset": 2048, 00:08:50.806 "data_size": 63488 00:08:50.806 }, 00:08:50.806 { 00:08:50.806 "name": "BaseBdev2", 00:08:50.806 "uuid": "43fddafa-ce11-4408-8b6e-1b19dfe9f9f3", 00:08:50.806 "is_configured": true, 00:08:50.806 "data_offset": 2048, 00:08:50.806 "data_size": 63488 00:08:50.806 }, 00:08:50.806 { 00:08:50.806 "name": "BaseBdev3", 00:08:50.806 "uuid": "07cc63b6-e3c4-4af3-acd3-84a4e8af7293", 00:08:50.806 "is_configured": true, 00:08:50.806 "data_offset": 2048, 00:08:50.806 "data_size": 63488 00:08:50.806 } 00:08:50.806 ] 00:08:50.806 }' 00:08:50.806 15:13:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:50.806 15:13:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:51.375 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:51.375 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:51.375 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:51.375 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:51.375 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:51.375 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:51.375 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:51.375 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:51.375 15:13:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.375 15:13:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:51.375 [2024-11-27 15:13:19.229006] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:51.375 15:13:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.375 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:51.375 "name": "Existed_Raid", 00:08:51.375 "aliases": [ 00:08:51.375 "070d3f19-b3f6-4750-bad0-4d82a0ee4ea4" 00:08:51.375 ], 00:08:51.375 "product_name": "Raid Volume", 00:08:51.375 "block_size": 512, 00:08:51.375 "num_blocks": 190464, 00:08:51.375 "uuid": "070d3f19-b3f6-4750-bad0-4d82a0ee4ea4", 00:08:51.375 "assigned_rate_limits": { 00:08:51.375 "rw_ios_per_sec": 0, 00:08:51.375 "rw_mbytes_per_sec": 0, 00:08:51.375 "r_mbytes_per_sec": 0, 00:08:51.375 "w_mbytes_per_sec": 0 00:08:51.375 }, 00:08:51.375 "claimed": false, 00:08:51.375 "zoned": false, 00:08:51.375 "supported_io_types": { 00:08:51.375 "read": true, 00:08:51.375 "write": true, 00:08:51.375 "unmap": true, 00:08:51.375 "flush": true, 00:08:51.375 "reset": true, 00:08:51.375 "nvme_admin": false, 00:08:51.375 "nvme_io": false, 00:08:51.375 "nvme_io_md": false, 00:08:51.375 "write_zeroes": true, 00:08:51.375 "zcopy": false, 00:08:51.375 "get_zone_info": false, 00:08:51.375 "zone_management": false, 00:08:51.375 "zone_append": false, 00:08:51.375 "compare": false, 00:08:51.375 "compare_and_write": false, 00:08:51.375 "abort": false, 00:08:51.375 "seek_hole": false, 00:08:51.375 "seek_data": false, 00:08:51.375 "copy": false, 00:08:51.375 "nvme_iov_md": false 00:08:51.375 }, 00:08:51.375 "memory_domains": [ 00:08:51.375 { 00:08:51.375 "dma_device_id": "system", 00:08:51.375 "dma_device_type": 1 00:08:51.375 }, 00:08:51.375 { 00:08:51.375 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:51.375 "dma_device_type": 2 00:08:51.375 }, 00:08:51.375 { 00:08:51.375 "dma_device_id": "system", 00:08:51.375 "dma_device_type": 1 00:08:51.375 }, 00:08:51.375 { 00:08:51.375 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:51.375 "dma_device_type": 2 00:08:51.375 }, 00:08:51.375 { 00:08:51.375 "dma_device_id": "system", 00:08:51.375 "dma_device_type": 1 00:08:51.375 }, 00:08:51.375 { 00:08:51.375 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:51.375 "dma_device_type": 2 00:08:51.375 } 00:08:51.375 ], 00:08:51.375 "driver_specific": { 00:08:51.375 "raid": { 00:08:51.375 "uuid": "070d3f19-b3f6-4750-bad0-4d82a0ee4ea4", 00:08:51.375 "strip_size_kb": 64, 00:08:51.375 "state": "online", 00:08:51.375 "raid_level": "concat", 00:08:51.375 "superblock": true, 00:08:51.375 "num_base_bdevs": 3, 00:08:51.375 "num_base_bdevs_discovered": 3, 00:08:51.375 "num_base_bdevs_operational": 3, 00:08:51.375 "base_bdevs_list": [ 00:08:51.375 { 00:08:51.375 "name": "BaseBdev1", 00:08:51.375 "uuid": "9f8d3326-62ce-4f94-80a7-e9a0c7f31cd9", 00:08:51.375 "is_configured": true, 00:08:51.375 "data_offset": 2048, 00:08:51.375 "data_size": 63488 00:08:51.375 }, 00:08:51.375 { 00:08:51.375 "name": "BaseBdev2", 00:08:51.375 "uuid": "43fddafa-ce11-4408-8b6e-1b19dfe9f9f3", 00:08:51.375 "is_configured": true, 00:08:51.375 "data_offset": 2048, 00:08:51.375 "data_size": 63488 00:08:51.375 }, 00:08:51.375 { 00:08:51.375 "name": "BaseBdev3", 00:08:51.376 "uuid": "07cc63b6-e3c4-4af3-acd3-84a4e8af7293", 00:08:51.376 "is_configured": true, 00:08:51.376 "data_offset": 2048, 00:08:51.376 "data_size": 63488 00:08:51.376 } 00:08:51.376 ] 00:08:51.376 } 00:08:51.376 } 00:08:51.376 }' 00:08:51.376 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:51.376 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:51.376 BaseBdev2 00:08:51.376 BaseBdev3' 00:08:51.376 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:51.376 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:51.376 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:51.376 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:51.376 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:51.376 15:13:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.376 15:13:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:51.376 15:13:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.376 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:51.376 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:51.376 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:51.376 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:51.376 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:51.376 15:13:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.376 15:13:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:51.376 15:13:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.376 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:51.376 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:51.376 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:51.376 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:51.376 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:51.376 15:13:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.376 15:13:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:51.376 15:13:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.636 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:51.636 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:51.636 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:51.636 15:13:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.636 15:13:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:51.636 [2024-11-27 15:13:19.508244] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:51.636 [2024-11-27 15:13:19.508314] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:51.636 [2024-11-27 15:13:19.508389] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:51.636 15:13:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.636 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:51.636 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:08:51.636 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:51.636 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:08:51.636 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:08:51.636 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 2 00:08:51.636 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:51.636 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:08:51.636 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:51.636 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:51.636 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:51.636 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:51.636 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:51.636 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:51.636 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:51.636 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:51.636 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:51.636 15:13:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.636 15:13:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:51.636 15:13:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.636 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:51.636 "name": "Existed_Raid", 00:08:51.636 "uuid": "070d3f19-b3f6-4750-bad0-4d82a0ee4ea4", 00:08:51.636 "strip_size_kb": 64, 00:08:51.636 "state": "offline", 00:08:51.636 "raid_level": "concat", 00:08:51.636 "superblock": true, 00:08:51.636 "num_base_bdevs": 3, 00:08:51.636 "num_base_bdevs_discovered": 2, 00:08:51.636 "num_base_bdevs_operational": 2, 00:08:51.636 "base_bdevs_list": [ 00:08:51.636 { 00:08:51.636 "name": null, 00:08:51.636 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:51.636 "is_configured": false, 00:08:51.636 "data_offset": 0, 00:08:51.636 "data_size": 63488 00:08:51.636 }, 00:08:51.636 { 00:08:51.636 "name": "BaseBdev2", 00:08:51.636 "uuid": "43fddafa-ce11-4408-8b6e-1b19dfe9f9f3", 00:08:51.636 "is_configured": true, 00:08:51.636 "data_offset": 2048, 00:08:51.636 "data_size": 63488 00:08:51.636 }, 00:08:51.636 { 00:08:51.636 "name": "BaseBdev3", 00:08:51.636 "uuid": "07cc63b6-e3c4-4af3-acd3-84a4e8af7293", 00:08:51.636 "is_configured": true, 00:08:51.636 "data_offset": 2048, 00:08:51.636 "data_size": 63488 00:08:51.636 } 00:08:51.636 ] 00:08:51.636 }' 00:08:51.636 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:51.636 15:13:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:51.896 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:51.896 15:13:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:52.156 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:52.156 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:52.156 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.156 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:52.156 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.156 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:52.156 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:52.156 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:52.156 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.156 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:52.156 [2024-11-27 15:13:20.034729] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:52.157 [2024-11-27 15:13:20.089878] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:52.157 [2024-11-27 15:13:20.090009] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:52.157 BaseBdev2 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:52.157 [ 00:08:52.157 { 00:08:52.157 "name": "BaseBdev2", 00:08:52.157 "aliases": [ 00:08:52.157 "d60af8d9-1294-4a18-a26a-a8f0dd17c740" 00:08:52.157 ], 00:08:52.157 "product_name": "Malloc disk", 00:08:52.157 "block_size": 512, 00:08:52.157 "num_blocks": 65536, 00:08:52.157 "uuid": "d60af8d9-1294-4a18-a26a-a8f0dd17c740", 00:08:52.157 "assigned_rate_limits": { 00:08:52.157 "rw_ios_per_sec": 0, 00:08:52.157 "rw_mbytes_per_sec": 0, 00:08:52.157 "r_mbytes_per_sec": 0, 00:08:52.157 "w_mbytes_per_sec": 0 00:08:52.157 }, 00:08:52.157 "claimed": false, 00:08:52.157 "zoned": false, 00:08:52.157 "supported_io_types": { 00:08:52.157 "read": true, 00:08:52.157 "write": true, 00:08:52.157 "unmap": true, 00:08:52.157 "flush": true, 00:08:52.157 "reset": true, 00:08:52.157 "nvme_admin": false, 00:08:52.157 "nvme_io": false, 00:08:52.157 "nvme_io_md": false, 00:08:52.157 "write_zeroes": true, 00:08:52.157 "zcopy": true, 00:08:52.157 "get_zone_info": false, 00:08:52.157 "zone_management": false, 00:08:52.157 "zone_append": false, 00:08:52.157 "compare": false, 00:08:52.157 "compare_and_write": false, 00:08:52.157 "abort": true, 00:08:52.157 "seek_hole": false, 00:08:52.157 "seek_data": false, 00:08:52.157 "copy": true, 00:08:52.157 "nvme_iov_md": false 00:08:52.157 }, 00:08:52.157 "memory_domains": [ 00:08:52.157 { 00:08:52.157 "dma_device_id": "system", 00:08:52.157 "dma_device_type": 1 00:08:52.157 }, 00:08:52.157 { 00:08:52.157 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:52.157 "dma_device_type": 2 00:08:52.157 } 00:08:52.157 ], 00:08:52.157 "driver_specific": {} 00:08:52.157 } 00:08:52.157 ] 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:52.157 BaseBdev3 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:52.157 [ 00:08:52.157 { 00:08:52.157 "name": "BaseBdev3", 00:08:52.157 "aliases": [ 00:08:52.157 "f3dc31d0-70bd-4561-af4a-b58968c8f09b" 00:08:52.157 ], 00:08:52.157 "product_name": "Malloc disk", 00:08:52.157 "block_size": 512, 00:08:52.157 "num_blocks": 65536, 00:08:52.157 "uuid": "f3dc31d0-70bd-4561-af4a-b58968c8f09b", 00:08:52.157 "assigned_rate_limits": { 00:08:52.157 "rw_ios_per_sec": 0, 00:08:52.157 "rw_mbytes_per_sec": 0, 00:08:52.157 "r_mbytes_per_sec": 0, 00:08:52.157 "w_mbytes_per_sec": 0 00:08:52.157 }, 00:08:52.157 "claimed": false, 00:08:52.157 "zoned": false, 00:08:52.157 "supported_io_types": { 00:08:52.157 "read": true, 00:08:52.157 "write": true, 00:08:52.157 "unmap": true, 00:08:52.157 "flush": true, 00:08:52.157 "reset": true, 00:08:52.157 "nvme_admin": false, 00:08:52.157 "nvme_io": false, 00:08:52.157 "nvme_io_md": false, 00:08:52.157 "write_zeroes": true, 00:08:52.157 "zcopy": true, 00:08:52.157 "get_zone_info": false, 00:08:52.157 "zone_management": false, 00:08:52.157 "zone_append": false, 00:08:52.157 "compare": false, 00:08:52.157 "compare_and_write": false, 00:08:52.157 "abort": true, 00:08:52.157 "seek_hole": false, 00:08:52.157 "seek_data": false, 00:08:52.157 "copy": true, 00:08:52.157 "nvme_iov_md": false 00:08:52.157 }, 00:08:52.157 "memory_domains": [ 00:08:52.157 { 00:08:52.157 "dma_device_id": "system", 00:08:52.157 "dma_device_type": 1 00:08:52.157 }, 00:08:52.157 { 00:08:52.157 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:52.157 "dma_device_type": 2 00:08:52.157 } 00:08:52.157 ], 00:08:52.157 "driver_specific": {} 00:08:52.157 } 00:08:52.157 ] 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:52.157 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:52.416 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:52.416 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.416 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:52.416 [2024-11-27 15:13:20.267079] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:52.416 [2024-11-27 15:13:20.267169] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:52.416 [2024-11-27 15:13:20.267230] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:52.416 [2024-11-27 15:13:20.269064] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:52.416 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.416 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:52.416 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:52.416 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:52.416 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:52.416 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:52.416 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:52.416 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:52.416 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:52.417 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:52.417 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:52.417 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:52.417 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:52.417 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.417 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:52.417 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.417 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:52.417 "name": "Existed_Raid", 00:08:52.417 "uuid": "d370ba3f-c818-4046-b3fd-26bba7052883", 00:08:52.417 "strip_size_kb": 64, 00:08:52.417 "state": "configuring", 00:08:52.417 "raid_level": "concat", 00:08:52.417 "superblock": true, 00:08:52.417 "num_base_bdevs": 3, 00:08:52.417 "num_base_bdevs_discovered": 2, 00:08:52.417 "num_base_bdevs_operational": 3, 00:08:52.417 "base_bdevs_list": [ 00:08:52.417 { 00:08:52.417 "name": "BaseBdev1", 00:08:52.417 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:52.417 "is_configured": false, 00:08:52.417 "data_offset": 0, 00:08:52.417 "data_size": 0 00:08:52.417 }, 00:08:52.417 { 00:08:52.417 "name": "BaseBdev2", 00:08:52.417 "uuid": "d60af8d9-1294-4a18-a26a-a8f0dd17c740", 00:08:52.417 "is_configured": true, 00:08:52.417 "data_offset": 2048, 00:08:52.417 "data_size": 63488 00:08:52.417 }, 00:08:52.417 { 00:08:52.417 "name": "BaseBdev3", 00:08:52.417 "uuid": "f3dc31d0-70bd-4561-af4a-b58968c8f09b", 00:08:52.417 "is_configured": true, 00:08:52.417 "data_offset": 2048, 00:08:52.417 "data_size": 63488 00:08:52.417 } 00:08:52.417 ] 00:08:52.417 }' 00:08:52.417 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:52.417 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:52.675 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:08:52.675 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.675 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:52.675 [2024-11-27 15:13:20.742313] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:52.675 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.675 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:52.675 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:52.675 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:52.675 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:52.675 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:52.675 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:52.675 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:52.675 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:52.675 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:52.675 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:52.675 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:52.675 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:52.675 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.675 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:52.675 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.934 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:52.934 "name": "Existed_Raid", 00:08:52.934 "uuid": "d370ba3f-c818-4046-b3fd-26bba7052883", 00:08:52.934 "strip_size_kb": 64, 00:08:52.934 "state": "configuring", 00:08:52.934 "raid_level": "concat", 00:08:52.934 "superblock": true, 00:08:52.934 "num_base_bdevs": 3, 00:08:52.934 "num_base_bdevs_discovered": 1, 00:08:52.934 "num_base_bdevs_operational": 3, 00:08:52.934 "base_bdevs_list": [ 00:08:52.934 { 00:08:52.934 "name": "BaseBdev1", 00:08:52.934 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:52.934 "is_configured": false, 00:08:52.934 "data_offset": 0, 00:08:52.934 "data_size": 0 00:08:52.934 }, 00:08:52.934 { 00:08:52.934 "name": null, 00:08:52.934 "uuid": "d60af8d9-1294-4a18-a26a-a8f0dd17c740", 00:08:52.934 "is_configured": false, 00:08:52.934 "data_offset": 0, 00:08:52.934 "data_size": 63488 00:08:52.934 }, 00:08:52.934 { 00:08:52.934 "name": "BaseBdev3", 00:08:52.934 "uuid": "f3dc31d0-70bd-4561-af4a-b58968c8f09b", 00:08:52.934 "is_configured": true, 00:08:52.934 "data_offset": 2048, 00:08:52.934 "data_size": 63488 00:08:52.934 } 00:08:52.934 ] 00:08:52.934 }' 00:08:52.934 15:13:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:52.934 15:13:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.194 [2024-11-27 15:13:21.176662] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:53.194 BaseBdev1 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.194 [ 00:08:53.194 { 00:08:53.194 "name": "BaseBdev1", 00:08:53.194 "aliases": [ 00:08:53.194 "86f00d09-a839-464d-9fab-a535bb5c0859" 00:08:53.194 ], 00:08:53.194 "product_name": "Malloc disk", 00:08:53.194 "block_size": 512, 00:08:53.194 "num_blocks": 65536, 00:08:53.194 "uuid": "86f00d09-a839-464d-9fab-a535bb5c0859", 00:08:53.194 "assigned_rate_limits": { 00:08:53.194 "rw_ios_per_sec": 0, 00:08:53.194 "rw_mbytes_per_sec": 0, 00:08:53.194 "r_mbytes_per_sec": 0, 00:08:53.194 "w_mbytes_per_sec": 0 00:08:53.194 }, 00:08:53.194 "claimed": true, 00:08:53.194 "claim_type": "exclusive_write", 00:08:53.194 "zoned": false, 00:08:53.194 "supported_io_types": { 00:08:53.194 "read": true, 00:08:53.194 "write": true, 00:08:53.194 "unmap": true, 00:08:53.194 "flush": true, 00:08:53.194 "reset": true, 00:08:53.194 "nvme_admin": false, 00:08:53.194 "nvme_io": false, 00:08:53.194 "nvme_io_md": false, 00:08:53.194 "write_zeroes": true, 00:08:53.194 "zcopy": true, 00:08:53.194 "get_zone_info": false, 00:08:53.194 "zone_management": false, 00:08:53.194 "zone_append": false, 00:08:53.194 "compare": false, 00:08:53.194 "compare_and_write": false, 00:08:53.194 "abort": true, 00:08:53.194 "seek_hole": false, 00:08:53.194 "seek_data": false, 00:08:53.194 "copy": true, 00:08:53.194 "nvme_iov_md": false 00:08:53.194 }, 00:08:53.194 "memory_domains": [ 00:08:53.194 { 00:08:53.194 "dma_device_id": "system", 00:08:53.194 "dma_device_type": 1 00:08:53.194 }, 00:08:53.194 { 00:08:53.194 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:53.194 "dma_device_type": 2 00:08:53.194 } 00:08:53.194 ], 00:08:53.194 "driver_specific": {} 00:08:53.194 } 00:08:53.194 ] 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:53.194 "name": "Existed_Raid", 00:08:53.194 "uuid": "d370ba3f-c818-4046-b3fd-26bba7052883", 00:08:53.194 "strip_size_kb": 64, 00:08:53.194 "state": "configuring", 00:08:53.194 "raid_level": "concat", 00:08:53.194 "superblock": true, 00:08:53.194 "num_base_bdevs": 3, 00:08:53.194 "num_base_bdevs_discovered": 2, 00:08:53.194 "num_base_bdevs_operational": 3, 00:08:53.194 "base_bdevs_list": [ 00:08:53.194 { 00:08:53.194 "name": "BaseBdev1", 00:08:53.194 "uuid": "86f00d09-a839-464d-9fab-a535bb5c0859", 00:08:53.194 "is_configured": true, 00:08:53.194 "data_offset": 2048, 00:08:53.194 "data_size": 63488 00:08:53.194 }, 00:08:53.194 { 00:08:53.194 "name": null, 00:08:53.194 "uuid": "d60af8d9-1294-4a18-a26a-a8f0dd17c740", 00:08:53.194 "is_configured": false, 00:08:53.194 "data_offset": 0, 00:08:53.194 "data_size": 63488 00:08:53.194 }, 00:08:53.194 { 00:08:53.194 "name": "BaseBdev3", 00:08:53.194 "uuid": "f3dc31d0-70bd-4561-af4a-b58968c8f09b", 00:08:53.194 "is_configured": true, 00:08:53.194 "data_offset": 2048, 00:08:53.194 "data_size": 63488 00:08:53.194 } 00:08:53.194 ] 00:08:53.194 }' 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:53.194 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.765 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:53.765 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:53.765 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.765 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.765 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.765 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:08:53.765 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:08:53.765 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.765 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.765 [2024-11-27 15:13:21.667927] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:53.765 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.765 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:53.765 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:53.765 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:53.765 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:53.765 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:53.765 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:53.765 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:53.765 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:53.765 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:53.765 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:53.765 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:53.765 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:53.765 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.765 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.765 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.765 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:53.765 "name": "Existed_Raid", 00:08:53.765 "uuid": "d370ba3f-c818-4046-b3fd-26bba7052883", 00:08:53.765 "strip_size_kb": 64, 00:08:53.765 "state": "configuring", 00:08:53.765 "raid_level": "concat", 00:08:53.765 "superblock": true, 00:08:53.765 "num_base_bdevs": 3, 00:08:53.765 "num_base_bdevs_discovered": 1, 00:08:53.765 "num_base_bdevs_operational": 3, 00:08:53.765 "base_bdevs_list": [ 00:08:53.765 { 00:08:53.765 "name": "BaseBdev1", 00:08:53.765 "uuid": "86f00d09-a839-464d-9fab-a535bb5c0859", 00:08:53.765 "is_configured": true, 00:08:53.765 "data_offset": 2048, 00:08:53.765 "data_size": 63488 00:08:53.765 }, 00:08:53.765 { 00:08:53.765 "name": null, 00:08:53.765 "uuid": "d60af8d9-1294-4a18-a26a-a8f0dd17c740", 00:08:53.765 "is_configured": false, 00:08:53.765 "data_offset": 0, 00:08:53.765 "data_size": 63488 00:08:53.765 }, 00:08:53.765 { 00:08:53.765 "name": null, 00:08:53.765 "uuid": "f3dc31d0-70bd-4561-af4a-b58968c8f09b", 00:08:53.765 "is_configured": false, 00:08:53.765 "data_offset": 0, 00:08:53.765 "data_size": 63488 00:08:53.765 } 00:08:53.765 ] 00:08:53.765 }' 00:08:53.765 15:13:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:53.765 15:13:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:54.025 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:54.025 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:54.025 15:13:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:54.025 15:13:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:54.285 15:13:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:54.285 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:08:54.285 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:08:54.285 15:13:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:54.285 15:13:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:54.285 [2024-11-27 15:13:22.155820] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:54.285 15:13:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:54.285 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:54.285 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:54.285 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:54.286 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:54.286 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:54.286 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:54.286 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:54.286 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:54.286 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:54.286 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:54.286 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:54.286 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:54.286 15:13:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:54.286 15:13:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:54.286 15:13:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:54.286 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:54.286 "name": "Existed_Raid", 00:08:54.286 "uuid": "d370ba3f-c818-4046-b3fd-26bba7052883", 00:08:54.286 "strip_size_kb": 64, 00:08:54.286 "state": "configuring", 00:08:54.286 "raid_level": "concat", 00:08:54.286 "superblock": true, 00:08:54.286 "num_base_bdevs": 3, 00:08:54.286 "num_base_bdevs_discovered": 2, 00:08:54.286 "num_base_bdevs_operational": 3, 00:08:54.286 "base_bdevs_list": [ 00:08:54.286 { 00:08:54.286 "name": "BaseBdev1", 00:08:54.286 "uuid": "86f00d09-a839-464d-9fab-a535bb5c0859", 00:08:54.286 "is_configured": true, 00:08:54.286 "data_offset": 2048, 00:08:54.286 "data_size": 63488 00:08:54.286 }, 00:08:54.286 { 00:08:54.286 "name": null, 00:08:54.286 "uuid": "d60af8d9-1294-4a18-a26a-a8f0dd17c740", 00:08:54.286 "is_configured": false, 00:08:54.286 "data_offset": 0, 00:08:54.286 "data_size": 63488 00:08:54.286 }, 00:08:54.286 { 00:08:54.286 "name": "BaseBdev3", 00:08:54.286 "uuid": "f3dc31d0-70bd-4561-af4a-b58968c8f09b", 00:08:54.286 "is_configured": true, 00:08:54.286 "data_offset": 2048, 00:08:54.286 "data_size": 63488 00:08:54.286 } 00:08:54.286 ] 00:08:54.286 }' 00:08:54.286 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:54.286 15:13:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:54.546 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:54.546 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:54.546 15:13:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:54.546 15:13:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:54.546 15:13:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:54.546 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:08:54.546 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:54.546 15:13:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:54.546 15:13:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:54.546 [2024-11-27 15:13:22.595868] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:54.546 15:13:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:54.546 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:54.546 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:54.546 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:54.546 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:54.546 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:54.546 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:54.546 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:54.546 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:54.546 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:54.546 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:54.546 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:54.546 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:54.546 15:13:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:54.546 15:13:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:54.546 15:13:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:54.546 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:54.546 "name": "Existed_Raid", 00:08:54.546 "uuid": "d370ba3f-c818-4046-b3fd-26bba7052883", 00:08:54.546 "strip_size_kb": 64, 00:08:54.546 "state": "configuring", 00:08:54.546 "raid_level": "concat", 00:08:54.546 "superblock": true, 00:08:54.546 "num_base_bdevs": 3, 00:08:54.546 "num_base_bdevs_discovered": 1, 00:08:54.546 "num_base_bdevs_operational": 3, 00:08:54.546 "base_bdevs_list": [ 00:08:54.546 { 00:08:54.546 "name": null, 00:08:54.546 "uuid": "86f00d09-a839-464d-9fab-a535bb5c0859", 00:08:54.546 "is_configured": false, 00:08:54.546 "data_offset": 0, 00:08:54.546 "data_size": 63488 00:08:54.546 }, 00:08:54.546 { 00:08:54.546 "name": null, 00:08:54.546 "uuid": "d60af8d9-1294-4a18-a26a-a8f0dd17c740", 00:08:54.546 "is_configured": false, 00:08:54.546 "data_offset": 0, 00:08:54.546 "data_size": 63488 00:08:54.546 }, 00:08:54.546 { 00:08:54.546 "name": "BaseBdev3", 00:08:54.546 "uuid": "f3dc31d0-70bd-4561-af4a-b58968c8f09b", 00:08:54.546 "is_configured": true, 00:08:54.546 "data_offset": 2048, 00:08:54.546 "data_size": 63488 00:08:54.546 } 00:08:54.546 ] 00:08:54.546 }' 00:08:54.546 15:13:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:54.546 15:13:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.115 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:55.115 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:55.115 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.115 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.115 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.115 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:08:55.115 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:08:55.115 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.115 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.115 [2024-11-27 15:13:23.113625] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:55.115 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.115 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:55.115 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:55.115 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:55.115 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:55.115 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:55.115 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:55.115 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:55.115 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:55.115 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:55.115 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:55.115 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:55.115 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:55.115 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.115 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.115 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.115 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:55.115 "name": "Existed_Raid", 00:08:55.115 "uuid": "d370ba3f-c818-4046-b3fd-26bba7052883", 00:08:55.115 "strip_size_kb": 64, 00:08:55.115 "state": "configuring", 00:08:55.115 "raid_level": "concat", 00:08:55.115 "superblock": true, 00:08:55.115 "num_base_bdevs": 3, 00:08:55.115 "num_base_bdevs_discovered": 2, 00:08:55.115 "num_base_bdevs_operational": 3, 00:08:55.115 "base_bdevs_list": [ 00:08:55.116 { 00:08:55.116 "name": null, 00:08:55.116 "uuid": "86f00d09-a839-464d-9fab-a535bb5c0859", 00:08:55.116 "is_configured": false, 00:08:55.116 "data_offset": 0, 00:08:55.116 "data_size": 63488 00:08:55.116 }, 00:08:55.116 { 00:08:55.116 "name": "BaseBdev2", 00:08:55.116 "uuid": "d60af8d9-1294-4a18-a26a-a8f0dd17c740", 00:08:55.116 "is_configured": true, 00:08:55.116 "data_offset": 2048, 00:08:55.116 "data_size": 63488 00:08:55.116 }, 00:08:55.116 { 00:08:55.116 "name": "BaseBdev3", 00:08:55.116 "uuid": "f3dc31d0-70bd-4561-af4a-b58968c8f09b", 00:08:55.116 "is_configured": true, 00:08:55.116 "data_offset": 2048, 00:08:55.116 "data_size": 63488 00:08:55.116 } 00:08:55.116 ] 00:08:55.116 }' 00:08:55.116 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:55.116 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.696 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:55.696 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:55.696 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.696 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.696 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.696 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:08:55.696 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:55.696 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:08:55.696 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.696 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.696 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.696 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 86f00d09-a839-464d-9fab-a535bb5c0859 00:08:55.696 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.696 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.696 [2024-11-27 15:13:23.671868] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:08:55.696 NewBaseBdev 00:08:55.696 [2024-11-27 15:13:23.672110] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:08:55.696 [2024-11-27 15:13:23.672132] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:55.696 [2024-11-27 15:13:23.672376] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:08:55.696 [2024-11-27 15:13:23.672493] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:08:55.696 [2024-11-27 15:13:23.672502] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006d00 00:08:55.696 [2024-11-27 15:13:23.672600] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:55.696 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.696 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:08:55.696 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:08:55.696 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:55.696 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:55.696 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:55.696 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:55.696 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:55.696 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.696 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.696 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.696 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:08:55.696 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.696 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.696 [ 00:08:55.696 { 00:08:55.696 "name": "NewBaseBdev", 00:08:55.696 "aliases": [ 00:08:55.696 "86f00d09-a839-464d-9fab-a535bb5c0859" 00:08:55.697 ], 00:08:55.697 "product_name": "Malloc disk", 00:08:55.697 "block_size": 512, 00:08:55.697 "num_blocks": 65536, 00:08:55.697 "uuid": "86f00d09-a839-464d-9fab-a535bb5c0859", 00:08:55.697 "assigned_rate_limits": { 00:08:55.697 "rw_ios_per_sec": 0, 00:08:55.697 "rw_mbytes_per_sec": 0, 00:08:55.697 "r_mbytes_per_sec": 0, 00:08:55.697 "w_mbytes_per_sec": 0 00:08:55.697 }, 00:08:55.697 "claimed": true, 00:08:55.697 "claim_type": "exclusive_write", 00:08:55.697 "zoned": false, 00:08:55.697 "supported_io_types": { 00:08:55.697 "read": true, 00:08:55.697 "write": true, 00:08:55.697 "unmap": true, 00:08:55.697 "flush": true, 00:08:55.697 "reset": true, 00:08:55.697 "nvme_admin": false, 00:08:55.697 "nvme_io": false, 00:08:55.697 "nvme_io_md": false, 00:08:55.697 "write_zeroes": true, 00:08:55.697 "zcopy": true, 00:08:55.697 "get_zone_info": false, 00:08:55.697 "zone_management": false, 00:08:55.697 "zone_append": false, 00:08:55.697 "compare": false, 00:08:55.697 "compare_and_write": false, 00:08:55.697 "abort": true, 00:08:55.697 "seek_hole": false, 00:08:55.697 "seek_data": false, 00:08:55.697 "copy": true, 00:08:55.697 "nvme_iov_md": false 00:08:55.697 }, 00:08:55.697 "memory_domains": [ 00:08:55.697 { 00:08:55.697 "dma_device_id": "system", 00:08:55.697 "dma_device_type": 1 00:08:55.697 }, 00:08:55.697 { 00:08:55.697 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:55.697 "dma_device_type": 2 00:08:55.697 } 00:08:55.697 ], 00:08:55.697 "driver_specific": {} 00:08:55.697 } 00:08:55.697 ] 00:08:55.697 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.697 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:55.697 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:08:55.697 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:55.697 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:55.697 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:55.697 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:55.697 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:55.697 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:55.697 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:55.697 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:55.697 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:55.697 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:55.697 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:55.697 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.697 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.697 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.697 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:55.697 "name": "Existed_Raid", 00:08:55.697 "uuid": "d370ba3f-c818-4046-b3fd-26bba7052883", 00:08:55.697 "strip_size_kb": 64, 00:08:55.697 "state": "online", 00:08:55.697 "raid_level": "concat", 00:08:55.697 "superblock": true, 00:08:55.697 "num_base_bdevs": 3, 00:08:55.697 "num_base_bdevs_discovered": 3, 00:08:55.697 "num_base_bdevs_operational": 3, 00:08:55.697 "base_bdevs_list": [ 00:08:55.697 { 00:08:55.697 "name": "NewBaseBdev", 00:08:55.697 "uuid": "86f00d09-a839-464d-9fab-a535bb5c0859", 00:08:55.697 "is_configured": true, 00:08:55.697 "data_offset": 2048, 00:08:55.697 "data_size": 63488 00:08:55.697 }, 00:08:55.697 { 00:08:55.697 "name": "BaseBdev2", 00:08:55.697 "uuid": "d60af8d9-1294-4a18-a26a-a8f0dd17c740", 00:08:55.697 "is_configured": true, 00:08:55.697 "data_offset": 2048, 00:08:55.697 "data_size": 63488 00:08:55.697 }, 00:08:55.697 { 00:08:55.697 "name": "BaseBdev3", 00:08:55.697 "uuid": "f3dc31d0-70bd-4561-af4a-b58968c8f09b", 00:08:55.697 "is_configured": true, 00:08:55.697 "data_offset": 2048, 00:08:55.697 "data_size": 63488 00:08:55.697 } 00:08:55.697 ] 00:08:55.697 }' 00:08:55.697 15:13:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:55.697 15:13:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:56.267 15:13:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:08:56.267 15:13:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:56.267 15:13:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:56.267 15:13:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:56.267 15:13:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:56.267 15:13:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:56.267 15:13:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:56.267 15:13:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:56.267 15:13:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:56.267 15:13:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:56.267 [2024-11-27 15:13:24.183387] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:56.267 15:13:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:56.267 15:13:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:56.267 "name": "Existed_Raid", 00:08:56.267 "aliases": [ 00:08:56.267 "d370ba3f-c818-4046-b3fd-26bba7052883" 00:08:56.267 ], 00:08:56.267 "product_name": "Raid Volume", 00:08:56.267 "block_size": 512, 00:08:56.267 "num_blocks": 190464, 00:08:56.267 "uuid": "d370ba3f-c818-4046-b3fd-26bba7052883", 00:08:56.267 "assigned_rate_limits": { 00:08:56.267 "rw_ios_per_sec": 0, 00:08:56.267 "rw_mbytes_per_sec": 0, 00:08:56.267 "r_mbytes_per_sec": 0, 00:08:56.267 "w_mbytes_per_sec": 0 00:08:56.267 }, 00:08:56.267 "claimed": false, 00:08:56.267 "zoned": false, 00:08:56.267 "supported_io_types": { 00:08:56.267 "read": true, 00:08:56.267 "write": true, 00:08:56.267 "unmap": true, 00:08:56.267 "flush": true, 00:08:56.267 "reset": true, 00:08:56.267 "nvme_admin": false, 00:08:56.267 "nvme_io": false, 00:08:56.267 "nvme_io_md": false, 00:08:56.267 "write_zeroes": true, 00:08:56.267 "zcopy": false, 00:08:56.267 "get_zone_info": false, 00:08:56.267 "zone_management": false, 00:08:56.267 "zone_append": false, 00:08:56.267 "compare": false, 00:08:56.267 "compare_and_write": false, 00:08:56.267 "abort": false, 00:08:56.267 "seek_hole": false, 00:08:56.267 "seek_data": false, 00:08:56.267 "copy": false, 00:08:56.267 "nvme_iov_md": false 00:08:56.267 }, 00:08:56.267 "memory_domains": [ 00:08:56.267 { 00:08:56.267 "dma_device_id": "system", 00:08:56.267 "dma_device_type": 1 00:08:56.267 }, 00:08:56.267 { 00:08:56.267 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:56.267 "dma_device_type": 2 00:08:56.267 }, 00:08:56.267 { 00:08:56.267 "dma_device_id": "system", 00:08:56.267 "dma_device_type": 1 00:08:56.267 }, 00:08:56.267 { 00:08:56.267 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:56.267 "dma_device_type": 2 00:08:56.267 }, 00:08:56.267 { 00:08:56.267 "dma_device_id": "system", 00:08:56.267 "dma_device_type": 1 00:08:56.267 }, 00:08:56.267 { 00:08:56.267 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:56.267 "dma_device_type": 2 00:08:56.267 } 00:08:56.267 ], 00:08:56.267 "driver_specific": { 00:08:56.267 "raid": { 00:08:56.267 "uuid": "d370ba3f-c818-4046-b3fd-26bba7052883", 00:08:56.267 "strip_size_kb": 64, 00:08:56.267 "state": "online", 00:08:56.267 "raid_level": "concat", 00:08:56.267 "superblock": true, 00:08:56.267 "num_base_bdevs": 3, 00:08:56.267 "num_base_bdevs_discovered": 3, 00:08:56.267 "num_base_bdevs_operational": 3, 00:08:56.267 "base_bdevs_list": [ 00:08:56.267 { 00:08:56.267 "name": "NewBaseBdev", 00:08:56.267 "uuid": "86f00d09-a839-464d-9fab-a535bb5c0859", 00:08:56.267 "is_configured": true, 00:08:56.267 "data_offset": 2048, 00:08:56.267 "data_size": 63488 00:08:56.267 }, 00:08:56.267 { 00:08:56.267 "name": "BaseBdev2", 00:08:56.267 "uuid": "d60af8d9-1294-4a18-a26a-a8f0dd17c740", 00:08:56.267 "is_configured": true, 00:08:56.267 "data_offset": 2048, 00:08:56.267 "data_size": 63488 00:08:56.267 }, 00:08:56.267 { 00:08:56.267 "name": "BaseBdev3", 00:08:56.267 "uuid": "f3dc31d0-70bd-4561-af4a-b58968c8f09b", 00:08:56.267 "is_configured": true, 00:08:56.267 "data_offset": 2048, 00:08:56.267 "data_size": 63488 00:08:56.267 } 00:08:56.267 ] 00:08:56.267 } 00:08:56.267 } 00:08:56.267 }' 00:08:56.267 15:13:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:56.267 15:13:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:08:56.267 BaseBdev2 00:08:56.267 BaseBdev3' 00:08:56.267 15:13:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:56.267 15:13:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:56.267 15:13:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:56.267 15:13:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:56.267 15:13:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:08:56.267 15:13:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:56.268 15:13:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:56.268 15:13:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:56.268 15:13:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:56.268 15:13:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:56.268 15:13:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:56.268 15:13:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:56.268 15:13:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:56.268 15:13:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:56.268 15:13:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:56.268 15:13:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:56.528 15:13:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:56.528 15:13:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:56.528 15:13:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:56.528 15:13:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:56.528 15:13:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:56.528 15:13:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:56.528 15:13:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:56.528 15:13:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:56.528 15:13:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:56.528 15:13:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:56.528 15:13:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:56.528 15:13:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:56.528 15:13:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:56.528 [2024-11-27 15:13:24.446594] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:56.528 [2024-11-27 15:13:24.446663] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:56.528 [2024-11-27 15:13:24.446751] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:56.528 [2024-11-27 15:13:24.446821] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:56.528 [2024-11-27 15:13:24.446874] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name Existed_Raid, state offline 00:08:56.528 15:13:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:56.528 15:13:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 77512 00:08:56.528 15:13:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 77512 ']' 00:08:56.528 15:13:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 77512 00:08:56.528 15:13:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:08:56.528 15:13:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:56.528 15:13:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 77512 00:08:56.528 15:13:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:56.528 15:13:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:56.528 15:13:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 77512' 00:08:56.528 killing process with pid 77512 00:08:56.528 15:13:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 77512 00:08:56.528 [2024-11-27 15:13:24.484517] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:56.528 15:13:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 77512 00:08:56.528 [2024-11-27 15:13:24.515745] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:56.788 15:13:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:08:56.788 00:08:56.788 real 0m8.943s 00:08:56.788 user 0m15.218s 00:08:56.788 sys 0m1.960s 00:08:56.788 ************************************ 00:08:56.788 END TEST raid_state_function_test_sb 00:08:56.788 ************************************ 00:08:56.788 15:13:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:56.788 15:13:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:56.788 15:13:24 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 3 00:08:56.788 15:13:24 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:08:56.788 15:13:24 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:56.788 15:13:24 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:56.788 ************************************ 00:08:56.788 START TEST raid_superblock_test 00:08:56.788 ************************************ 00:08:56.788 15:13:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test concat 3 00:08:56.788 15:13:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:08:56.788 15:13:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:08:56.788 15:13:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:08:56.788 15:13:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:08:56.788 15:13:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:08:56.788 15:13:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:08:56.788 15:13:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:08:56.788 15:13:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:08:56.788 15:13:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:08:56.788 15:13:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:08:56.788 15:13:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:08:56.788 15:13:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:08:56.788 15:13:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:08:56.788 15:13:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:08:56.788 15:13:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:08:56.788 15:13:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:08:56.788 15:13:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=78121 00:08:56.788 15:13:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:08:56.788 15:13:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 78121 00:08:56.788 15:13:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 78121 ']' 00:08:56.788 15:13:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:56.788 15:13:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:56.788 15:13:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:56.788 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:56.788 15:13:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:56.788 15:13:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.048 [2024-11-27 15:13:24.900688] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:08:57.048 [2024-11-27 15:13:24.900817] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid78121 ] 00:08:57.048 [2024-11-27 15:13:25.072805] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:57.048 [2024-11-27 15:13:25.099022] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:57.048 [2024-11-27 15:13:25.142083] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:57.048 [2024-11-27 15:13:25.142166] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.987 malloc1 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.987 [2024-11-27 15:13:25.750689] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:57.987 [2024-11-27 15:13:25.750804] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:57.987 [2024-11-27 15:13:25.750866] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:08:57.987 [2024-11-27 15:13:25.750919] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:57.987 [2024-11-27 15:13:25.753021] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:57.987 [2024-11-27 15:13:25.753099] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:57.987 pt1 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.987 malloc2 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.987 [2024-11-27 15:13:25.779463] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:57.987 [2024-11-27 15:13:25.779586] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:57.987 [2024-11-27 15:13:25.779620] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:08:57.987 [2024-11-27 15:13:25.779671] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:57.987 [2024-11-27 15:13:25.781775] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:57.987 [2024-11-27 15:13:25.781848] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:57.987 pt2 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.987 malloc3 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.987 15:13:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.987 [2024-11-27 15:13:25.808100] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:08:57.987 [2024-11-27 15:13:25.808188] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:57.987 [2024-11-27 15:13:25.808238] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:08:57.987 [2024-11-27 15:13:25.808268] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:57.987 [2024-11-27 15:13:25.810246] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:57.988 [2024-11-27 15:13:25.810317] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:08:57.988 pt3 00:08:57.988 15:13:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.988 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:57.988 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:57.988 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:08:57.988 15:13:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.988 15:13:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.988 [2024-11-27 15:13:25.820117] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:57.988 [2024-11-27 15:13:25.821873] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:57.988 [2024-11-27 15:13:25.821974] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:08:57.988 [2024-11-27 15:13:25.822128] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:08:57.988 [2024-11-27 15:13:25.822193] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:57.988 [2024-11-27 15:13:25.822471] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:08:57.988 [2024-11-27 15:13:25.822633] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:08:57.988 [2024-11-27 15:13:25.822676] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:08:57.988 [2024-11-27 15:13:25.822832] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:57.988 15:13:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.988 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:08:57.988 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:57.988 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:57.988 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:57.988 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:57.988 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:57.988 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:57.988 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:57.988 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:57.988 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:57.988 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:57.988 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:57.988 15:13:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.988 15:13:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.988 15:13:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.988 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:57.988 "name": "raid_bdev1", 00:08:57.988 "uuid": "17f84ba4-aefb-42a1-b388-e9683b249a2f", 00:08:57.988 "strip_size_kb": 64, 00:08:57.988 "state": "online", 00:08:57.988 "raid_level": "concat", 00:08:57.988 "superblock": true, 00:08:57.988 "num_base_bdevs": 3, 00:08:57.988 "num_base_bdevs_discovered": 3, 00:08:57.988 "num_base_bdevs_operational": 3, 00:08:57.988 "base_bdevs_list": [ 00:08:57.988 { 00:08:57.988 "name": "pt1", 00:08:57.988 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:57.988 "is_configured": true, 00:08:57.988 "data_offset": 2048, 00:08:57.988 "data_size": 63488 00:08:57.988 }, 00:08:57.988 { 00:08:57.988 "name": "pt2", 00:08:57.988 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:57.988 "is_configured": true, 00:08:57.988 "data_offset": 2048, 00:08:57.988 "data_size": 63488 00:08:57.988 }, 00:08:57.988 { 00:08:57.988 "name": "pt3", 00:08:57.988 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:57.988 "is_configured": true, 00:08:57.988 "data_offset": 2048, 00:08:57.988 "data_size": 63488 00:08:57.988 } 00:08:57.988 ] 00:08:57.988 }' 00:08:57.988 15:13:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:57.988 15:13:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.248 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:08:58.248 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:58.248 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:58.248 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:58.248 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:58.248 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:58.248 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:58.248 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:58.248 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.248 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.248 [2024-11-27 15:13:26.231918] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:58.248 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.248 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:58.248 "name": "raid_bdev1", 00:08:58.248 "aliases": [ 00:08:58.248 "17f84ba4-aefb-42a1-b388-e9683b249a2f" 00:08:58.248 ], 00:08:58.248 "product_name": "Raid Volume", 00:08:58.248 "block_size": 512, 00:08:58.248 "num_blocks": 190464, 00:08:58.248 "uuid": "17f84ba4-aefb-42a1-b388-e9683b249a2f", 00:08:58.248 "assigned_rate_limits": { 00:08:58.248 "rw_ios_per_sec": 0, 00:08:58.248 "rw_mbytes_per_sec": 0, 00:08:58.248 "r_mbytes_per_sec": 0, 00:08:58.248 "w_mbytes_per_sec": 0 00:08:58.248 }, 00:08:58.248 "claimed": false, 00:08:58.248 "zoned": false, 00:08:58.248 "supported_io_types": { 00:08:58.248 "read": true, 00:08:58.248 "write": true, 00:08:58.248 "unmap": true, 00:08:58.248 "flush": true, 00:08:58.248 "reset": true, 00:08:58.248 "nvme_admin": false, 00:08:58.248 "nvme_io": false, 00:08:58.248 "nvme_io_md": false, 00:08:58.248 "write_zeroes": true, 00:08:58.248 "zcopy": false, 00:08:58.248 "get_zone_info": false, 00:08:58.248 "zone_management": false, 00:08:58.248 "zone_append": false, 00:08:58.248 "compare": false, 00:08:58.248 "compare_and_write": false, 00:08:58.248 "abort": false, 00:08:58.248 "seek_hole": false, 00:08:58.248 "seek_data": false, 00:08:58.248 "copy": false, 00:08:58.248 "nvme_iov_md": false 00:08:58.248 }, 00:08:58.248 "memory_domains": [ 00:08:58.248 { 00:08:58.248 "dma_device_id": "system", 00:08:58.248 "dma_device_type": 1 00:08:58.248 }, 00:08:58.248 { 00:08:58.248 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:58.248 "dma_device_type": 2 00:08:58.248 }, 00:08:58.248 { 00:08:58.248 "dma_device_id": "system", 00:08:58.248 "dma_device_type": 1 00:08:58.248 }, 00:08:58.248 { 00:08:58.248 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:58.248 "dma_device_type": 2 00:08:58.248 }, 00:08:58.248 { 00:08:58.248 "dma_device_id": "system", 00:08:58.248 "dma_device_type": 1 00:08:58.248 }, 00:08:58.248 { 00:08:58.248 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:58.248 "dma_device_type": 2 00:08:58.248 } 00:08:58.248 ], 00:08:58.248 "driver_specific": { 00:08:58.248 "raid": { 00:08:58.248 "uuid": "17f84ba4-aefb-42a1-b388-e9683b249a2f", 00:08:58.248 "strip_size_kb": 64, 00:08:58.248 "state": "online", 00:08:58.248 "raid_level": "concat", 00:08:58.248 "superblock": true, 00:08:58.248 "num_base_bdevs": 3, 00:08:58.248 "num_base_bdevs_discovered": 3, 00:08:58.248 "num_base_bdevs_operational": 3, 00:08:58.248 "base_bdevs_list": [ 00:08:58.248 { 00:08:58.248 "name": "pt1", 00:08:58.248 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:58.248 "is_configured": true, 00:08:58.248 "data_offset": 2048, 00:08:58.248 "data_size": 63488 00:08:58.248 }, 00:08:58.248 { 00:08:58.248 "name": "pt2", 00:08:58.248 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:58.248 "is_configured": true, 00:08:58.248 "data_offset": 2048, 00:08:58.248 "data_size": 63488 00:08:58.248 }, 00:08:58.248 { 00:08:58.248 "name": "pt3", 00:08:58.248 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:58.248 "is_configured": true, 00:08:58.248 "data_offset": 2048, 00:08:58.248 "data_size": 63488 00:08:58.248 } 00:08:58.248 ] 00:08:58.248 } 00:08:58.248 } 00:08:58.248 }' 00:08:58.248 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:58.249 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:58.249 pt2 00:08:58.249 pt3' 00:08:58.249 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.509 [2024-11-27 15:13:26.519310] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=17f84ba4-aefb-42a1-b388-e9683b249a2f 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 17f84ba4-aefb-42a1-b388-e9683b249a2f ']' 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.509 [2024-11-27 15:13:26.563017] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:58.509 [2024-11-27 15:13:26.563081] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:58.509 [2024-11-27 15:13:26.563195] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:58.509 [2024-11-27 15:13:26.563287] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:58.509 [2024-11-27 15:13:26.563342] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:08:58.509 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.769 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:08:58.769 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:08:58.769 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:58.769 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:08:58.769 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.769 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.769 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.769 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:58.769 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:08:58.769 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.769 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.769 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.769 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:58.769 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:08:58.769 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.769 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.769 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.769 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:08:58.769 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:08:58.769 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.769 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.769 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.769 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:08:58.769 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:58.769 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:08:58.769 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:58.769 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:08:58.769 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:58.769 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:08:58.769 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:58.769 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:58.769 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.770 [2024-11-27 15:13:26.718834] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:08:58.770 [2024-11-27 15:13:26.720823] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:08:58.770 [2024-11-27 15:13:26.720876] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:08:58.770 [2024-11-27 15:13:26.720942] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:08:58.770 [2024-11-27 15:13:26.720997] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:08:58.770 [2024-11-27 15:13:26.721019] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:08:58.770 [2024-11-27 15:13:26.721034] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:58.770 [2024-11-27 15:13:26.721045] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state configuring 00:08:58.770 request: 00:08:58.770 { 00:08:58.770 "name": "raid_bdev1", 00:08:58.770 "raid_level": "concat", 00:08:58.770 "base_bdevs": [ 00:08:58.770 "malloc1", 00:08:58.770 "malloc2", 00:08:58.770 "malloc3" 00:08:58.770 ], 00:08:58.770 "strip_size_kb": 64, 00:08:58.770 "superblock": false, 00:08:58.770 "method": "bdev_raid_create", 00:08:58.770 "req_id": 1 00:08:58.770 } 00:08:58.770 Got JSON-RPC error response 00:08:58.770 response: 00:08:58.770 { 00:08:58.770 "code": -17, 00:08:58.770 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:08:58.770 } 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.770 [2024-11-27 15:13:26.778674] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:58.770 [2024-11-27 15:13:26.778732] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:58.770 [2024-11-27 15:13:26.778749] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:08:58.770 [2024-11-27 15:13:26.778760] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:58.770 [2024-11-27 15:13:26.780876] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:58.770 [2024-11-27 15:13:26.780929] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:58.770 [2024-11-27 15:13:26.781001] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:08:58.770 [2024-11-27 15:13:26.781048] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:58.770 pt1 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 3 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:58.770 "name": "raid_bdev1", 00:08:58.770 "uuid": "17f84ba4-aefb-42a1-b388-e9683b249a2f", 00:08:58.770 "strip_size_kb": 64, 00:08:58.770 "state": "configuring", 00:08:58.770 "raid_level": "concat", 00:08:58.770 "superblock": true, 00:08:58.770 "num_base_bdevs": 3, 00:08:58.770 "num_base_bdevs_discovered": 1, 00:08:58.770 "num_base_bdevs_operational": 3, 00:08:58.770 "base_bdevs_list": [ 00:08:58.770 { 00:08:58.770 "name": "pt1", 00:08:58.770 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:58.770 "is_configured": true, 00:08:58.770 "data_offset": 2048, 00:08:58.770 "data_size": 63488 00:08:58.770 }, 00:08:58.770 { 00:08:58.770 "name": null, 00:08:58.770 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:58.770 "is_configured": false, 00:08:58.770 "data_offset": 2048, 00:08:58.770 "data_size": 63488 00:08:58.770 }, 00:08:58.770 { 00:08:58.770 "name": null, 00:08:58.770 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:58.770 "is_configured": false, 00:08:58.770 "data_offset": 2048, 00:08:58.770 "data_size": 63488 00:08:58.770 } 00:08:58.770 ] 00:08:58.770 }' 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:58.770 15:13:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.340 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:08:59.340 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:59.340 15:13:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:59.340 15:13:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.340 [2024-11-27 15:13:27.245940] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:59.340 [2024-11-27 15:13:27.246004] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:59.340 [2024-11-27 15:13:27.246024] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:08:59.340 [2024-11-27 15:13:27.246037] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:59.340 [2024-11-27 15:13:27.246464] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:59.340 [2024-11-27 15:13:27.246494] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:59.340 [2024-11-27 15:13:27.246574] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:59.340 [2024-11-27 15:13:27.246609] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:59.340 pt2 00:08:59.340 15:13:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:59.340 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:08:59.340 15:13:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:59.340 15:13:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.340 [2024-11-27 15:13:27.257881] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:08:59.340 15:13:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:59.340 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 3 00:08:59.340 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:59.340 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:59.340 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:59.340 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:59.340 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:59.340 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:59.340 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:59.340 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:59.340 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:59.340 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:59.340 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:59.340 15:13:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:59.340 15:13:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.340 15:13:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:59.340 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:59.340 "name": "raid_bdev1", 00:08:59.340 "uuid": "17f84ba4-aefb-42a1-b388-e9683b249a2f", 00:08:59.340 "strip_size_kb": 64, 00:08:59.340 "state": "configuring", 00:08:59.340 "raid_level": "concat", 00:08:59.340 "superblock": true, 00:08:59.340 "num_base_bdevs": 3, 00:08:59.340 "num_base_bdevs_discovered": 1, 00:08:59.340 "num_base_bdevs_operational": 3, 00:08:59.340 "base_bdevs_list": [ 00:08:59.340 { 00:08:59.340 "name": "pt1", 00:08:59.340 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:59.340 "is_configured": true, 00:08:59.340 "data_offset": 2048, 00:08:59.340 "data_size": 63488 00:08:59.340 }, 00:08:59.340 { 00:08:59.340 "name": null, 00:08:59.340 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:59.340 "is_configured": false, 00:08:59.340 "data_offset": 0, 00:08:59.340 "data_size": 63488 00:08:59.340 }, 00:08:59.340 { 00:08:59.340 "name": null, 00:08:59.340 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:59.340 "is_configured": false, 00:08:59.340 "data_offset": 2048, 00:08:59.340 "data_size": 63488 00:08:59.340 } 00:08:59.340 ] 00:08:59.340 }' 00:08:59.340 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:59.340 15:13:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.908 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:08:59.908 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:59.908 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:59.908 15:13:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:59.908 15:13:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.908 [2024-11-27 15:13:27.741062] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:59.908 [2024-11-27 15:13:27.741146] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:59.908 [2024-11-27 15:13:27.741170] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:08:59.908 [2024-11-27 15:13:27.741179] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:59.908 [2024-11-27 15:13:27.741583] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:59.908 [2024-11-27 15:13:27.741608] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:59.908 [2024-11-27 15:13:27.741685] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:59.908 [2024-11-27 15:13:27.741713] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:59.908 pt2 00:08:59.908 15:13:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:59.908 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:59.908 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:59.908 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:08:59.908 15:13:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:59.908 15:13:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.908 [2024-11-27 15:13:27.753033] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:08:59.908 [2024-11-27 15:13:27.753081] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:59.908 [2024-11-27 15:13:27.753100] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:08:59.908 [2024-11-27 15:13:27.753110] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:59.908 [2024-11-27 15:13:27.753444] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:59.908 [2024-11-27 15:13:27.753468] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:08:59.908 [2024-11-27 15:13:27.753530] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:08:59.908 [2024-11-27 15:13:27.753561] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:08:59.908 [2024-11-27 15:13:27.753665] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:08:59.908 [2024-11-27 15:13:27.753681] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:59.908 [2024-11-27 15:13:27.753937] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:08:59.908 [2024-11-27 15:13:27.754050] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:08:59.908 [2024-11-27 15:13:27.754067] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:08:59.908 [2024-11-27 15:13:27.754165] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:59.908 pt3 00:08:59.908 15:13:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:59.908 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:59.908 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:59.908 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:08:59.908 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:59.908 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:59.908 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:59.908 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:59.908 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:59.908 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:59.908 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:59.908 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:59.908 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:59.909 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:59.909 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:59.909 15:13:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:59.909 15:13:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.909 15:13:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:59.909 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:59.909 "name": "raid_bdev1", 00:08:59.909 "uuid": "17f84ba4-aefb-42a1-b388-e9683b249a2f", 00:08:59.909 "strip_size_kb": 64, 00:08:59.909 "state": "online", 00:08:59.909 "raid_level": "concat", 00:08:59.909 "superblock": true, 00:08:59.909 "num_base_bdevs": 3, 00:08:59.909 "num_base_bdevs_discovered": 3, 00:08:59.909 "num_base_bdevs_operational": 3, 00:08:59.909 "base_bdevs_list": [ 00:08:59.909 { 00:08:59.909 "name": "pt1", 00:08:59.909 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:59.909 "is_configured": true, 00:08:59.909 "data_offset": 2048, 00:08:59.909 "data_size": 63488 00:08:59.909 }, 00:08:59.909 { 00:08:59.909 "name": "pt2", 00:08:59.909 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:59.909 "is_configured": true, 00:08:59.909 "data_offset": 2048, 00:08:59.909 "data_size": 63488 00:08:59.909 }, 00:08:59.909 { 00:08:59.909 "name": "pt3", 00:08:59.909 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:59.909 "is_configured": true, 00:08:59.909 "data_offset": 2048, 00:08:59.909 "data_size": 63488 00:08:59.909 } 00:08:59.909 ] 00:08:59.909 }' 00:08:59.909 15:13:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:59.909 15:13:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:00.168 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:09:00.168 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:00.168 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:00.168 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:00.168 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:00.168 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:00.168 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:00.168 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:00.168 15:13:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.168 15:13:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:00.168 [2024-11-27 15:13:28.192572] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:00.168 15:13:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.168 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:00.168 "name": "raid_bdev1", 00:09:00.168 "aliases": [ 00:09:00.168 "17f84ba4-aefb-42a1-b388-e9683b249a2f" 00:09:00.168 ], 00:09:00.168 "product_name": "Raid Volume", 00:09:00.168 "block_size": 512, 00:09:00.168 "num_blocks": 190464, 00:09:00.168 "uuid": "17f84ba4-aefb-42a1-b388-e9683b249a2f", 00:09:00.168 "assigned_rate_limits": { 00:09:00.168 "rw_ios_per_sec": 0, 00:09:00.168 "rw_mbytes_per_sec": 0, 00:09:00.168 "r_mbytes_per_sec": 0, 00:09:00.168 "w_mbytes_per_sec": 0 00:09:00.168 }, 00:09:00.168 "claimed": false, 00:09:00.168 "zoned": false, 00:09:00.168 "supported_io_types": { 00:09:00.168 "read": true, 00:09:00.168 "write": true, 00:09:00.168 "unmap": true, 00:09:00.168 "flush": true, 00:09:00.168 "reset": true, 00:09:00.168 "nvme_admin": false, 00:09:00.168 "nvme_io": false, 00:09:00.168 "nvme_io_md": false, 00:09:00.168 "write_zeroes": true, 00:09:00.168 "zcopy": false, 00:09:00.168 "get_zone_info": false, 00:09:00.168 "zone_management": false, 00:09:00.168 "zone_append": false, 00:09:00.168 "compare": false, 00:09:00.168 "compare_and_write": false, 00:09:00.168 "abort": false, 00:09:00.168 "seek_hole": false, 00:09:00.168 "seek_data": false, 00:09:00.168 "copy": false, 00:09:00.168 "nvme_iov_md": false 00:09:00.168 }, 00:09:00.168 "memory_domains": [ 00:09:00.168 { 00:09:00.168 "dma_device_id": "system", 00:09:00.168 "dma_device_type": 1 00:09:00.168 }, 00:09:00.168 { 00:09:00.168 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:00.168 "dma_device_type": 2 00:09:00.168 }, 00:09:00.168 { 00:09:00.168 "dma_device_id": "system", 00:09:00.168 "dma_device_type": 1 00:09:00.168 }, 00:09:00.168 { 00:09:00.168 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:00.168 "dma_device_type": 2 00:09:00.168 }, 00:09:00.168 { 00:09:00.168 "dma_device_id": "system", 00:09:00.168 "dma_device_type": 1 00:09:00.168 }, 00:09:00.168 { 00:09:00.168 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:00.168 "dma_device_type": 2 00:09:00.168 } 00:09:00.168 ], 00:09:00.168 "driver_specific": { 00:09:00.168 "raid": { 00:09:00.168 "uuid": "17f84ba4-aefb-42a1-b388-e9683b249a2f", 00:09:00.168 "strip_size_kb": 64, 00:09:00.168 "state": "online", 00:09:00.168 "raid_level": "concat", 00:09:00.168 "superblock": true, 00:09:00.168 "num_base_bdevs": 3, 00:09:00.168 "num_base_bdevs_discovered": 3, 00:09:00.168 "num_base_bdevs_operational": 3, 00:09:00.168 "base_bdevs_list": [ 00:09:00.168 { 00:09:00.168 "name": "pt1", 00:09:00.168 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:00.168 "is_configured": true, 00:09:00.168 "data_offset": 2048, 00:09:00.168 "data_size": 63488 00:09:00.168 }, 00:09:00.168 { 00:09:00.168 "name": "pt2", 00:09:00.168 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:00.168 "is_configured": true, 00:09:00.168 "data_offset": 2048, 00:09:00.168 "data_size": 63488 00:09:00.168 }, 00:09:00.168 { 00:09:00.168 "name": "pt3", 00:09:00.168 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:00.168 "is_configured": true, 00:09:00.168 "data_offset": 2048, 00:09:00.168 "data_size": 63488 00:09:00.168 } 00:09:00.168 ] 00:09:00.168 } 00:09:00.168 } 00:09:00.168 }' 00:09:00.168 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:00.168 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:00.168 pt2 00:09:00.168 pt3' 00:09:00.168 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:00.427 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:00.427 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:00.427 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:00.427 15:13:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.427 15:13:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:00.427 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:00.427 15:13:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.427 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:00.427 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:00.427 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:00.427 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:00.427 15:13:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.427 15:13:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:00.427 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:00.427 15:13:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.427 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:00.427 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:00.427 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:00.428 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:00.428 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:00.428 15:13:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.428 15:13:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:00.428 15:13:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.428 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:00.428 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:00.428 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:00.428 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:09:00.428 15:13:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.428 15:13:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:00.428 [2024-11-27 15:13:28.460106] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:00.428 15:13:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.428 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 17f84ba4-aefb-42a1-b388-e9683b249a2f '!=' 17f84ba4-aefb-42a1-b388-e9683b249a2f ']' 00:09:00.428 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:09:00.428 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:00.428 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:00.428 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 78121 00:09:00.428 15:13:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 78121 ']' 00:09:00.428 15:13:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 78121 00:09:00.428 15:13:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:09:00.428 15:13:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:00.428 15:13:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 78121 00:09:00.687 15:13:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:00.687 15:13:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:00.687 killing process with pid 78121 00:09:00.687 15:13:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 78121' 00:09:00.687 15:13:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 78121 00:09:00.687 [2024-11-27 15:13:28.535845] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:00.687 [2024-11-27 15:13:28.535959] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:00.687 [2024-11-27 15:13:28.536026] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:00.687 [2024-11-27 15:13:28.536042] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:09:00.687 15:13:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 78121 00:09:00.687 [2024-11-27 15:13:28.569599] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:00.687 15:13:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:09:00.687 00:09:00.687 real 0m3.976s 00:09:00.687 user 0m6.292s 00:09:00.687 sys 0m0.858s 00:09:00.687 15:13:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:00.687 15:13:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:00.687 ************************************ 00:09:00.687 END TEST raid_superblock_test 00:09:00.687 ************************************ 00:09:00.947 15:13:28 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 3 read 00:09:00.947 15:13:28 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:00.947 15:13:28 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:00.947 15:13:28 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:00.947 ************************************ 00:09:00.947 START TEST raid_read_error_test 00:09:00.947 ************************************ 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 3 read 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.iI6B2963hD 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=78363 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 78363 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 78363 ']' 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:00.947 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:00.947 15:13:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:00.948 15:13:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:00.948 [2024-11-27 15:13:28.972393] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:09:00.948 [2024-11-27 15:13:28.972544] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid78363 ] 00:09:01.207 [2024-11-27 15:13:29.147454] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:01.207 [2024-11-27 15:13:29.173695] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:01.207 [2024-11-27 15:13:29.217465] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:01.207 [2024-11-27 15:13:29.217505] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.775 BaseBdev1_malloc 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.775 true 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.775 [2024-11-27 15:13:29.806073] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:01.775 [2024-11-27 15:13:29.806132] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:01.775 [2024-11-27 15:13:29.806152] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:09:01.775 [2024-11-27 15:13:29.806161] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:01.775 [2024-11-27 15:13:29.808285] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:01.775 [2024-11-27 15:13:29.808320] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:01.775 BaseBdev1 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.775 BaseBdev2_malloc 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.775 true 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.775 [2024-11-27 15:13:29.846740] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:01.775 [2024-11-27 15:13:29.846787] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:01.775 [2024-11-27 15:13:29.846806] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:09:01.775 [2024-11-27 15:13:29.846815] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:01.775 [2024-11-27 15:13:29.848857] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:01.775 [2024-11-27 15:13:29.848908] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:01.775 BaseBdev2 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.775 BaseBdev3_malloc 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.775 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.034 true 00:09:02.034 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.034 15:13:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:02.034 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.034 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.034 [2024-11-27 15:13:29.887478] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:02.034 [2024-11-27 15:13:29.887530] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:02.034 [2024-11-27 15:13:29.887550] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:09:02.034 [2024-11-27 15:13:29.887559] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:02.034 [2024-11-27 15:13:29.889573] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:02.034 [2024-11-27 15:13:29.889613] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:02.034 BaseBdev3 00:09:02.034 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.034 15:13:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:09:02.034 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.034 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.034 [2024-11-27 15:13:29.899519] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:02.034 [2024-11-27 15:13:29.901332] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:02.034 [2024-11-27 15:13:29.901414] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:02.034 [2024-11-27 15:13:29.901578] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:09:02.034 [2024-11-27 15:13:29.901606] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:09:02.034 [2024-11-27 15:13:29.901841] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:09:02.034 [2024-11-27 15:13:29.902023] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:09:02.034 [2024-11-27 15:13:29.902040] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006d00 00:09:02.034 [2024-11-27 15:13:29.902149] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:02.034 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.034 15:13:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:09:02.034 15:13:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:02.034 15:13:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:02.034 15:13:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:02.034 15:13:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:02.034 15:13:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:02.034 15:13:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:02.034 15:13:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:02.034 15:13:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:02.034 15:13:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:02.034 15:13:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:02.034 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.034 15:13:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:02.034 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.034 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.034 15:13:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:02.035 "name": "raid_bdev1", 00:09:02.035 "uuid": "235efc0d-f214-4705-b5eb-352a640e8577", 00:09:02.035 "strip_size_kb": 64, 00:09:02.035 "state": "online", 00:09:02.035 "raid_level": "concat", 00:09:02.035 "superblock": true, 00:09:02.035 "num_base_bdevs": 3, 00:09:02.035 "num_base_bdevs_discovered": 3, 00:09:02.035 "num_base_bdevs_operational": 3, 00:09:02.035 "base_bdevs_list": [ 00:09:02.035 { 00:09:02.035 "name": "BaseBdev1", 00:09:02.035 "uuid": "099baff6-318d-5774-ac8f-4865a650b970", 00:09:02.035 "is_configured": true, 00:09:02.035 "data_offset": 2048, 00:09:02.035 "data_size": 63488 00:09:02.035 }, 00:09:02.035 { 00:09:02.035 "name": "BaseBdev2", 00:09:02.035 "uuid": "81757dd7-704f-5a3b-9954-e1df13bf743a", 00:09:02.035 "is_configured": true, 00:09:02.035 "data_offset": 2048, 00:09:02.035 "data_size": 63488 00:09:02.035 }, 00:09:02.035 { 00:09:02.035 "name": "BaseBdev3", 00:09:02.035 "uuid": "3eda043f-7121-5a73-a9fc-5f089c3685e4", 00:09:02.035 "is_configured": true, 00:09:02.035 "data_offset": 2048, 00:09:02.035 "data_size": 63488 00:09:02.035 } 00:09:02.035 ] 00:09:02.035 }' 00:09:02.035 15:13:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:02.035 15:13:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.293 15:13:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:02.293 15:13:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:02.552 [2024-11-27 15:13:30.423044] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006560 00:09:03.489 15:13:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:09:03.489 15:13:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:03.489 15:13:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.489 15:13:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:03.489 15:13:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:03.489 15:13:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:09:03.489 15:13:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:09:03.489 15:13:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:09:03.489 15:13:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:03.489 15:13:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:03.489 15:13:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:03.489 15:13:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:03.489 15:13:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:03.489 15:13:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:03.489 15:13:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:03.489 15:13:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:03.490 15:13:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:03.490 15:13:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:03.490 15:13:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:03.490 15:13:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.490 15:13:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:03.490 15:13:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:03.490 15:13:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:03.490 "name": "raid_bdev1", 00:09:03.490 "uuid": "235efc0d-f214-4705-b5eb-352a640e8577", 00:09:03.490 "strip_size_kb": 64, 00:09:03.490 "state": "online", 00:09:03.490 "raid_level": "concat", 00:09:03.490 "superblock": true, 00:09:03.490 "num_base_bdevs": 3, 00:09:03.490 "num_base_bdevs_discovered": 3, 00:09:03.490 "num_base_bdevs_operational": 3, 00:09:03.490 "base_bdevs_list": [ 00:09:03.490 { 00:09:03.490 "name": "BaseBdev1", 00:09:03.490 "uuid": "099baff6-318d-5774-ac8f-4865a650b970", 00:09:03.490 "is_configured": true, 00:09:03.490 "data_offset": 2048, 00:09:03.490 "data_size": 63488 00:09:03.490 }, 00:09:03.490 { 00:09:03.490 "name": "BaseBdev2", 00:09:03.490 "uuid": "81757dd7-704f-5a3b-9954-e1df13bf743a", 00:09:03.490 "is_configured": true, 00:09:03.490 "data_offset": 2048, 00:09:03.490 "data_size": 63488 00:09:03.490 }, 00:09:03.490 { 00:09:03.490 "name": "BaseBdev3", 00:09:03.490 "uuid": "3eda043f-7121-5a73-a9fc-5f089c3685e4", 00:09:03.490 "is_configured": true, 00:09:03.490 "data_offset": 2048, 00:09:03.490 "data_size": 63488 00:09:03.490 } 00:09:03.490 ] 00:09:03.490 }' 00:09:03.490 15:13:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:03.490 15:13:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.749 15:13:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:03.749 15:13:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:03.749 15:13:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.749 [2024-11-27 15:13:31.794728] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:03.749 [2024-11-27 15:13:31.794773] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:03.749 [2024-11-27 15:13:31.797458] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:03.749 [2024-11-27 15:13:31.797513] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:03.749 [2024-11-27 15:13:31.797550] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:03.749 [2024-11-27 15:13:31.797561] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name raid_bdev1, state offline 00:09:03.749 { 00:09:03.749 "results": [ 00:09:03.749 { 00:09:03.749 "job": "raid_bdev1", 00:09:03.749 "core_mask": "0x1", 00:09:03.749 "workload": "randrw", 00:09:03.749 "percentage": 50, 00:09:03.749 "status": "finished", 00:09:03.749 "queue_depth": 1, 00:09:03.749 "io_size": 131072, 00:09:03.749 "runtime": 1.372621, 00:09:03.749 "iops": 16255.761787121135, 00:09:03.749 "mibps": 2031.9702233901419, 00:09:03.749 "io_failed": 1, 00:09:03.749 "io_timeout": 0, 00:09:03.749 "avg_latency_us": 84.95567628837009, 00:09:03.749 "min_latency_us": 24.705676855895195, 00:09:03.749 "max_latency_us": 1438.071615720524 00:09:03.749 } 00:09:03.749 ], 00:09:03.749 "core_count": 1 00:09:03.749 } 00:09:03.749 15:13:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:03.749 15:13:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 78363 00:09:03.749 15:13:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 78363 ']' 00:09:03.749 15:13:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 78363 00:09:03.749 15:13:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:09:03.749 15:13:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:03.749 15:13:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 78363 00:09:03.749 15:13:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:03.749 15:13:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:03.749 killing process with pid 78363 00:09:03.749 15:13:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 78363' 00:09:03.749 15:13:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 78363 00:09:03.749 [2024-11-27 15:13:31.838683] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:03.749 15:13:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 78363 00:09:04.008 [2024-11-27 15:13:31.864818] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:04.008 15:13:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.iI6B2963hD 00:09:04.008 15:13:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:04.008 15:13:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:04.008 15:13:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.73 00:09:04.008 15:13:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:09:04.008 15:13:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:04.008 15:13:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:04.008 15:13:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.73 != \0\.\0\0 ]] 00:09:04.008 00:09:04.008 real 0m3.224s 00:09:04.008 user 0m4.069s 00:09:04.008 sys 0m0.539s 00:09:04.008 15:13:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:04.008 15:13:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.008 ************************************ 00:09:04.008 END TEST raid_read_error_test 00:09:04.008 ************************************ 00:09:04.274 15:13:32 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 3 write 00:09:04.274 15:13:32 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:04.274 15:13:32 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:04.274 15:13:32 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:04.274 ************************************ 00:09:04.274 START TEST raid_write_error_test 00:09:04.274 ************************************ 00:09:04.274 15:13:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 3 write 00:09:04.274 15:13:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:09:04.274 15:13:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:09:04.274 15:13:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:09:04.274 15:13:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:04.274 15:13:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:04.274 15:13:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:04.274 15:13:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:04.274 15:13:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:04.274 15:13:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:04.275 15:13:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:04.275 15:13:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:04.275 15:13:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:04.275 15:13:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:04.275 15:13:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:04.275 15:13:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:04.275 15:13:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:04.275 15:13:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:04.275 15:13:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:04.275 15:13:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:04.275 15:13:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:04.275 15:13:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:04.275 15:13:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:09:04.275 15:13:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:09:04.275 15:13:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:09:04.275 15:13:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:04.275 15:13:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.BihwvWoPVb 00:09:04.275 15:13:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=78492 00:09:04.275 15:13:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 78492 00:09:04.275 15:13:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:04.275 15:13:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 78492 ']' 00:09:04.275 15:13:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:04.275 15:13:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:04.275 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:04.275 15:13:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:04.275 15:13:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:04.275 15:13:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.275 [2024-11-27 15:13:32.261555] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:09:04.275 [2024-11-27 15:13:32.261670] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid78492 ] 00:09:04.536 [2024-11-27 15:13:32.421620] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:04.536 [2024-11-27 15:13:32.448430] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:04.536 [2024-11-27 15:13:32.492985] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:04.536 [2024-11-27 15:13:32.493026] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:05.105 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:05.105 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:09:05.105 15:13:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:05.105 15:13:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:05.105 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.105 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.105 BaseBdev1_malloc 00:09:05.105 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.105 15:13:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:05.105 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.105 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.105 true 00:09:05.105 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.105 15:13:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:05.105 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.105 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.105 [2024-11-27 15:13:33.153790] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:05.105 [2024-11-27 15:13:33.153844] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:05.105 [2024-11-27 15:13:33.153865] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:09:05.105 [2024-11-27 15:13:33.153874] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:05.105 [2024-11-27 15:13:33.156012] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:05.105 [2024-11-27 15:13:33.156046] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:05.105 BaseBdev1 00:09:05.105 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.105 15:13:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:05.105 15:13:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:05.105 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.105 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.105 BaseBdev2_malloc 00:09:05.105 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.105 15:13:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:05.105 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.105 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.105 true 00:09:05.105 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.105 15:13:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:05.105 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.105 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.105 [2024-11-27 15:13:33.194629] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:05.105 [2024-11-27 15:13:33.194685] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:05.105 [2024-11-27 15:13:33.194703] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:09:05.105 [2024-11-27 15:13:33.194712] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:05.105 [2024-11-27 15:13:33.196781] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:05.105 [2024-11-27 15:13:33.196887] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:05.105 BaseBdev2 00:09:05.105 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.105 15:13:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:05.105 15:13:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:05.105 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.105 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.365 BaseBdev3_malloc 00:09:05.365 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.365 15:13:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:05.365 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.365 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.365 true 00:09:05.365 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.365 15:13:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:05.365 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.365 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.365 [2024-11-27 15:13:33.235506] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:05.365 [2024-11-27 15:13:33.235560] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:05.365 [2024-11-27 15:13:33.235578] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:09:05.365 [2024-11-27 15:13:33.235586] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:05.365 [2024-11-27 15:13:33.237664] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:05.365 [2024-11-27 15:13:33.237703] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:05.365 BaseBdev3 00:09:05.365 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.365 15:13:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:09:05.365 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.365 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.365 [2024-11-27 15:13:33.247561] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:05.365 [2024-11-27 15:13:33.249505] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:05.365 [2024-11-27 15:13:33.249668] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:05.366 [2024-11-27 15:13:33.249855] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:09:05.366 [2024-11-27 15:13:33.249870] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:09:05.366 [2024-11-27 15:13:33.250162] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:09:05.366 [2024-11-27 15:13:33.250328] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:09:05.366 [2024-11-27 15:13:33.250340] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006d00 00:09:05.366 [2024-11-27 15:13:33.250489] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:05.366 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.366 15:13:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:09:05.366 15:13:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:05.366 15:13:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:05.366 15:13:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:05.366 15:13:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:05.366 15:13:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:05.366 15:13:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:05.366 15:13:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:05.366 15:13:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:05.366 15:13:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:05.366 15:13:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:05.366 15:13:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:05.366 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.366 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.366 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.366 15:13:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:05.366 "name": "raid_bdev1", 00:09:05.366 "uuid": "20225c92-8ba5-4960-b855-e036638c4f85", 00:09:05.366 "strip_size_kb": 64, 00:09:05.366 "state": "online", 00:09:05.366 "raid_level": "concat", 00:09:05.366 "superblock": true, 00:09:05.366 "num_base_bdevs": 3, 00:09:05.366 "num_base_bdevs_discovered": 3, 00:09:05.366 "num_base_bdevs_operational": 3, 00:09:05.366 "base_bdevs_list": [ 00:09:05.366 { 00:09:05.366 "name": "BaseBdev1", 00:09:05.366 "uuid": "b023eb39-4097-50f5-a34d-b722084fbb56", 00:09:05.366 "is_configured": true, 00:09:05.366 "data_offset": 2048, 00:09:05.366 "data_size": 63488 00:09:05.366 }, 00:09:05.366 { 00:09:05.366 "name": "BaseBdev2", 00:09:05.366 "uuid": "4bc0361b-fd3f-5507-910a-82bbca7dd497", 00:09:05.366 "is_configured": true, 00:09:05.366 "data_offset": 2048, 00:09:05.366 "data_size": 63488 00:09:05.366 }, 00:09:05.366 { 00:09:05.366 "name": "BaseBdev3", 00:09:05.366 "uuid": "af291786-eb59-5467-8610-31b1295737f8", 00:09:05.366 "is_configured": true, 00:09:05.366 "data_offset": 2048, 00:09:05.366 "data_size": 63488 00:09:05.366 } 00:09:05.366 ] 00:09:05.366 }' 00:09:05.366 15:13:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:05.366 15:13:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.625 15:13:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:05.625 15:13:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:05.884 [2024-11-27 15:13:33.794985] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006560 00:09:06.821 15:13:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:09:06.821 15:13:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:06.821 15:13:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.821 15:13:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:06.821 15:13:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:06.821 15:13:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:09:06.821 15:13:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:09:06.821 15:13:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:09:06.821 15:13:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:06.821 15:13:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:06.821 15:13:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:06.821 15:13:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:06.821 15:13:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:06.821 15:13:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:06.821 15:13:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:06.821 15:13:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:06.821 15:13:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:06.821 15:13:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:06.821 15:13:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:06.821 15:13:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:06.821 15:13:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.821 15:13:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:06.821 15:13:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:06.821 "name": "raid_bdev1", 00:09:06.821 "uuid": "20225c92-8ba5-4960-b855-e036638c4f85", 00:09:06.821 "strip_size_kb": 64, 00:09:06.821 "state": "online", 00:09:06.821 "raid_level": "concat", 00:09:06.821 "superblock": true, 00:09:06.821 "num_base_bdevs": 3, 00:09:06.821 "num_base_bdevs_discovered": 3, 00:09:06.821 "num_base_bdevs_operational": 3, 00:09:06.821 "base_bdevs_list": [ 00:09:06.821 { 00:09:06.821 "name": "BaseBdev1", 00:09:06.821 "uuid": "b023eb39-4097-50f5-a34d-b722084fbb56", 00:09:06.821 "is_configured": true, 00:09:06.821 "data_offset": 2048, 00:09:06.821 "data_size": 63488 00:09:06.821 }, 00:09:06.821 { 00:09:06.821 "name": "BaseBdev2", 00:09:06.821 "uuid": "4bc0361b-fd3f-5507-910a-82bbca7dd497", 00:09:06.821 "is_configured": true, 00:09:06.821 "data_offset": 2048, 00:09:06.821 "data_size": 63488 00:09:06.821 }, 00:09:06.821 { 00:09:06.821 "name": "BaseBdev3", 00:09:06.821 "uuid": "af291786-eb59-5467-8610-31b1295737f8", 00:09:06.821 "is_configured": true, 00:09:06.821 "data_offset": 2048, 00:09:06.821 "data_size": 63488 00:09:06.821 } 00:09:06.821 ] 00:09:06.821 }' 00:09:06.821 15:13:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:06.821 15:13:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.390 15:13:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:07.390 15:13:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.390 15:13:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.390 [2024-11-27 15:13:35.195143] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:07.390 [2024-11-27 15:13:35.195193] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:07.390 [2024-11-27 15:13:35.197791] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:07.390 [2024-11-27 15:13:35.197841] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:07.390 [2024-11-27 15:13:35.197876] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:07.390 [2024-11-27 15:13:35.197887] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name raid_bdev1, state offline 00:09:07.390 { 00:09:07.390 "results": [ 00:09:07.390 { 00:09:07.390 "job": "raid_bdev1", 00:09:07.390 "core_mask": "0x1", 00:09:07.390 "workload": "randrw", 00:09:07.390 "percentage": 50, 00:09:07.390 "status": "finished", 00:09:07.390 "queue_depth": 1, 00:09:07.390 "io_size": 131072, 00:09:07.390 "runtime": 1.401057, 00:09:07.390 "iops": 16041.460126176165, 00:09:07.390 "mibps": 2005.1825157720207, 00:09:07.390 "io_failed": 1, 00:09:07.390 "io_timeout": 0, 00:09:07.390 "avg_latency_us": 86.07276093043642, 00:09:07.390 "min_latency_us": 27.053275109170304, 00:09:07.390 "max_latency_us": 1416.6078602620087 00:09:07.390 } 00:09:07.390 ], 00:09:07.390 "core_count": 1 00:09:07.390 } 00:09:07.390 15:13:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.390 15:13:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 78492 00:09:07.390 15:13:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 78492 ']' 00:09:07.390 15:13:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 78492 00:09:07.390 15:13:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:09:07.390 15:13:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:07.390 15:13:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 78492 00:09:07.390 15:13:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:07.390 15:13:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:07.390 15:13:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 78492' 00:09:07.390 killing process with pid 78492 00:09:07.390 15:13:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 78492 00:09:07.390 [2024-11-27 15:13:35.250527] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:07.390 15:13:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 78492 00:09:07.390 [2024-11-27 15:13:35.277077] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:07.390 15:13:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.BihwvWoPVb 00:09:07.390 15:13:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:07.390 15:13:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:07.650 15:13:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.71 00:09:07.650 15:13:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:09:07.650 15:13:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:07.650 15:13:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:07.650 15:13:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.71 != \0\.\0\0 ]] 00:09:07.650 00:09:07.650 real 0m3.346s 00:09:07.650 user 0m4.251s 00:09:07.650 sys 0m0.571s 00:09:07.650 15:13:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:07.650 ************************************ 00:09:07.650 END TEST raid_write_error_test 00:09:07.650 ************************************ 00:09:07.650 15:13:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.650 15:13:35 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:09:07.650 15:13:35 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 3 false 00:09:07.650 15:13:35 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:07.650 15:13:35 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:07.650 15:13:35 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:07.650 ************************************ 00:09:07.650 START TEST raid_state_function_test 00:09:07.650 ************************************ 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 3 false 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=78619 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 78619' 00:09:07.650 Process raid pid: 78619 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 78619 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 78619 ']' 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:07.650 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:07.650 15:13:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.650 [2024-11-27 15:13:35.673359] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:09:07.650 [2024-11-27 15:13:35.673479] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:07.910 [2024-11-27 15:13:35.845861] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:07.910 [2024-11-27 15:13:35.872438] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:07.910 [2024-11-27 15:13:35.916828] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:07.910 [2024-11-27 15:13:35.916973] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:08.478 15:13:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:08.478 15:13:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:09:08.478 15:13:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:08.478 15:13:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.478 15:13:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.478 [2024-11-27 15:13:36.552771] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:08.478 [2024-11-27 15:13:36.552855] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:08.478 [2024-11-27 15:13:36.552866] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:08.478 [2024-11-27 15:13:36.552877] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:08.478 [2024-11-27 15:13:36.552883] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:08.478 [2024-11-27 15:13:36.552893] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:08.478 15:13:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.478 15:13:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:08.478 15:13:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:08.478 15:13:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:08.478 15:13:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:08.478 15:13:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:08.478 15:13:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:08.478 15:13:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:08.478 15:13:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:08.478 15:13:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:08.478 15:13:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:08.478 15:13:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:08.478 15:13:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.478 15:13:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.478 15:13:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:08.478 15:13:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.738 15:13:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:08.738 "name": "Existed_Raid", 00:09:08.738 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:08.738 "strip_size_kb": 0, 00:09:08.738 "state": "configuring", 00:09:08.738 "raid_level": "raid1", 00:09:08.738 "superblock": false, 00:09:08.738 "num_base_bdevs": 3, 00:09:08.738 "num_base_bdevs_discovered": 0, 00:09:08.738 "num_base_bdevs_operational": 3, 00:09:08.738 "base_bdevs_list": [ 00:09:08.738 { 00:09:08.738 "name": "BaseBdev1", 00:09:08.738 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:08.738 "is_configured": false, 00:09:08.738 "data_offset": 0, 00:09:08.738 "data_size": 0 00:09:08.738 }, 00:09:08.738 { 00:09:08.738 "name": "BaseBdev2", 00:09:08.738 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:08.738 "is_configured": false, 00:09:08.738 "data_offset": 0, 00:09:08.738 "data_size": 0 00:09:08.738 }, 00:09:08.738 { 00:09:08.738 "name": "BaseBdev3", 00:09:08.738 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:08.738 "is_configured": false, 00:09:08.738 "data_offset": 0, 00:09:08.738 "data_size": 0 00:09:08.738 } 00:09:08.738 ] 00:09:08.738 }' 00:09:08.738 15:13:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:08.738 15:13:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.998 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:08.998 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.998 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.998 [2024-11-27 15:13:37.039883] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:08.998 [2024-11-27 15:13:37.040014] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:09:08.998 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.998 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:08.998 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.998 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.998 [2024-11-27 15:13:37.051853] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:08.998 [2024-11-27 15:13:37.051948] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:08.998 [2024-11-27 15:13:37.051977] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:08.998 [2024-11-27 15:13:37.052001] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:08.998 [2024-11-27 15:13:37.052020] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:08.998 [2024-11-27 15:13:37.052041] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:08.998 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.998 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:08.998 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.998 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.998 [2024-11-27 15:13:37.072882] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:08.998 BaseBdev1 00:09:08.998 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.998 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:08.998 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:08.998 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:08.998 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:08.998 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:08.998 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:08.998 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:08.998 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.999 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.999 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.999 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:08.999 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.999 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.999 [ 00:09:08.999 { 00:09:08.999 "name": "BaseBdev1", 00:09:08.999 "aliases": [ 00:09:08.999 "492e46d2-9fca-44dc-8ca4-7664a9004c96" 00:09:09.259 ], 00:09:09.259 "product_name": "Malloc disk", 00:09:09.259 "block_size": 512, 00:09:09.259 "num_blocks": 65536, 00:09:09.259 "uuid": "492e46d2-9fca-44dc-8ca4-7664a9004c96", 00:09:09.259 "assigned_rate_limits": { 00:09:09.259 "rw_ios_per_sec": 0, 00:09:09.259 "rw_mbytes_per_sec": 0, 00:09:09.259 "r_mbytes_per_sec": 0, 00:09:09.259 "w_mbytes_per_sec": 0 00:09:09.259 }, 00:09:09.259 "claimed": true, 00:09:09.259 "claim_type": "exclusive_write", 00:09:09.259 "zoned": false, 00:09:09.259 "supported_io_types": { 00:09:09.259 "read": true, 00:09:09.259 "write": true, 00:09:09.259 "unmap": true, 00:09:09.259 "flush": true, 00:09:09.259 "reset": true, 00:09:09.259 "nvme_admin": false, 00:09:09.259 "nvme_io": false, 00:09:09.259 "nvme_io_md": false, 00:09:09.259 "write_zeroes": true, 00:09:09.259 "zcopy": true, 00:09:09.259 "get_zone_info": false, 00:09:09.259 "zone_management": false, 00:09:09.259 "zone_append": false, 00:09:09.259 "compare": false, 00:09:09.259 "compare_and_write": false, 00:09:09.259 "abort": true, 00:09:09.259 "seek_hole": false, 00:09:09.259 "seek_data": false, 00:09:09.259 "copy": true, 00:09:09.259 "nvme_iov_md": false 00:09:09.259 }, 00:09:09.259 "memory_domains": [ 00:09:09.259 { 00:09:09.259 "dma_device_id": "system", 00:09:09.259 "dma_device_type": 1 00:09:09.259 }, 00:09:09.259 { 00:09:09.259 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:09.259 "dma_device_type": 2 00:09:09.259 } 00:09:09.259 ], 00:09:09.259 "driver_specific": {} 00:09:09.259 } 00:09:09.259 ] 00:09:09.259 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:09.259 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:09.259 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:09.259 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:09.259 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:09.259 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:09.259 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:09.259 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:09.259 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:09.259 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:09.259 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:09.259 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:09.259 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:09.259 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:09.259 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:09.259 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.259 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:09.259 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:09.259 "name": "Existed_Raid", 00:09:09.259 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:09.259 "strip_size_kb": 0, 00:09:09.259 "state": "configuring", 00:09:09.259 "raid_level": "raid1", 00:09:09.259 "superblock": false, 00:09:09.259 "num_base_bdevs": 3, 00:09:09.259 "num_base_bdevs_discovered": 1, 00:09:09.259 "num_base_bdevs_operational": 3, 00:09:09.259 "base_bdevs_list": [ 00:09:09.259 { 00:09:09.259 "name": "BaseBdev1", 00:09:09.259 "uuid": "492e46d2-9fca-44dc-8ca4-7664a9004c96", 00:09:09.259 "is_configured": true, 00:09:09.259 "data_offset": 0, 00:09:09.259 "data_size": 65536 00:09:09.259 }, 00:09:09.259 { 00:09:09.259 "name": "BaseBdev2", 00:09:09.259 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:09.259 "is_configured": false, 00:09:09.259 "data_offset": 0, 00:09:09.259 "data_size": 0 00:09:09.259 }, 00:09:09.259 { 00:09:09.259 "name": "BaseBdev3", 00:09:09.259 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:09.259 "is_configured": false, 00:09:09.259 "data_offset": 0, 00:09:09.259 "data_size": 0 00:09:09.259 } 00:09:09.259 ] 00:09:09.259 }' 00:09:09.259 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:09.259 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.519 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:09.519 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:09.519 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.519 [2024-11-27 15:13:37.480192] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:09.519 [2024-11-27 15:13:37.480245] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:09:09.519 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:09.519 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:09.519 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:09.519 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.519 [2024-11-27 15:13:37.492202] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:09.519 [2024-11-27 15:13:37.494023] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:09.519 [2024-11-27 15:13:37.494066] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:09.519 [2024-11-27 15:13:37.494075] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:09.519 [2024-11-27 15:13:37.494085] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:09.519 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:09.519 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:09.519 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:09.519 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:09.519 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:09.519 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:09.519 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:09.519 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:09.519 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:09.519 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:09.519 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:09.519 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:09.519 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:09.519 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:09.519 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:09.519 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.519 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:09.519 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:09.519 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:09.519 "name": "Existed_Raid", 00:09:09.519 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:09.519 "strip_size_kb": 0, 00:09:09.519 "state": "configuring", 00:09:09.519 "raid_level": "raid1", 00:09:09.519 "superblock": false, 00:09:09.519 "num_base_bdevs": 3, 00:09:09.519 "num_base_bdevs_discovered": 1, 00:09:09.519 "num_base_bdevs_operational": 3, 00:09:09.519 "base_bdevs_list": [ 00:09:09.519 { 00:09:09.519 "name": "BaseBdev1", 00:09:09.519 "uuid": "492e46d2-9fca-44dc-8ca4-7664a9004c96", 00:09:09.519 "is_configured": true, 00:09:09.519 "data_offset": 0, 00:09:09.519 "data_size": 65536 00:09:09.519 }, 00:09:09.519 { 00:09:09.519 "name": "BaseBdev2", 00:09:09.519 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:09.519 "is_configured": false, 00:09:09.519 "data_offset": 0, 00:09:09.519 "data_size": 0 00:09:09.519 }, 00:09:09.519 { 00:09:09.519 "name": "BaseBdev3", 00:09:09.519 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:09.519 "is_configured": false, 00:09:09.519 "data_offset": 0, 00:09:09.519 "data_size": 0 00:09:09.519 } 00:09:09.519 ] 00:09:09.519 }' 00:09:09.519 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:09.519 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.089 [2024-11-27 15:13:37.914644] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:10.089 BaseBdev2 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.089 [ 00:09:10.089 { 00:09:10.089 "name": "BaseBdev2", 00:09:10.089 "aliases": [ 00:09:10.089 "0056499c-2eba-43bf-b9e2-348f9d8144af" 00:09:10.089 ], 00:09:10.089 "product_name": "Malloc disk", 00:09:10.089 "block_size": 512, 00:09:10.089 "num_blocks": 65536, 00:09:10.089 "uuid": "0056499c-2eba-43bf-b9e2-348f9d8144af", 00:09:10.089 "assigned_rate_limits": { 00:09:10.089 "rw_ios_per_sec": 0, 00:09:10.089 "rw_mbytes_per_sec": 0, 00:09:10.089 "r_mbytes_per_sec": 0, 00:09:10.089 "w_mbytes_per_sec": 0 00:09:10.089 }, 00:09:10.089 "claimed": true, 00:09:10.089 "claim_type": "exclusive_write", 00:09:10.089 "zoned": false, 00:09:10.089 "supported_io_types": { 00:09:10.089 "read": true, 00:09:10.089 "write": true, 00:09:10.089 "unmap": true, 00:09:10.089 "flush": true, 00:09:10.089 "reset": true, 00:09:10.089 "nvme_admin": false, 00:09:10.089 "nvme_io": false, 00:09:10.089 "nvme_io_md": false, 00:09:10.089 "write_zeroes": true, 00:09:10.089 "zcopy": true, 00:09:10.089 "get_zone_info": false, 00:09:10.089 "zone_management": false, 00:09:10.089 "zone_append": false, 00:09:10.089 "compare": false, 00:09:10.089 "compare_and_write": false, 00:09:10.089 "abort": true, 00:09:10.089 "seek_hole": false, 00:09:10.089 "seek_data": false, 00:09:10.089 "copy": true, 00:09:10.089 "nvme_iov_md": false 00:09:10.089 }, 00:09:10.089 "memory_domains": [ 00:09:10.089 { 00:09:10.089 "dma_device_id": "system", 00:09:10.089 "dma_device_type": 1 00:09:10.089 }, 00:09:10.089 { 00:09:10.089 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:10.089 "dma_device_type": 2 00:09:10.089 } 00:09:10.089 ], 00:09:10.089 "driver_specific": {} 00:09:10.089 } 00:09:10.089 ] 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.089 15:13:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:10.089 15:13:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:10.089 "name": "Existed_Raid", 00:09:10.089 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:10.089 "strip_size_kb": 0, 00:09:10.089 "state": "configuring", 00:09:10.089 "raid_level": "raid1", 00:09:10.089 "superblock": false, 00:09:10.089 "num_base_bdevs": 3, 00:09:10.089 "num_base_bdevs_discovered": 2, 00:09:10.089 "num_base_bdevs_operational": 3, 00:09:10.089 "base_bdevs_list": [ 00:09:10.089 { 00:09:10.089 "name": "BaseBdev1", 00:09:10.089 "uuid": "492e46d2-9fca-44dc-8ca4-7664a9004c96", 00:09:10.089 "is_configured": true, 00:09:10.089 "data_offset": 0, 00:09:10.089 "data_size": 65536 00:09:10.089 }, 00:09:10.089 { 00:09:10.089 "name": "BaseBdev2", 00:09:10.089 "uuid": "0056499c-2eba-43bf-b9e2-348f9d8144af", 00:09:10.089 "is_configured": true, 00:09:10.089 "data_offset": 0, 00:09:10.089 "data_size": 65536 00:09:10.090 }, 00:09:10.090 { 00:09:10.090 "name": "BaseBdev3", 00:09:10.090 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:10.090 "is_configured": false, 00:09:10.090 "data_offset": 0, 00:09:10.090 "data_size": 0 00:09:10.090 } 00:09:10.090 ] 00:09:10.090 }' 00:09:10.090 15:13:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:10.090 15:13:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.349 15:13:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:10.349 15:13:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:10.349 15:13:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.349 [2024-11-27 15:13:38.415838] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:10.349 [2024-11-27 15:13:38.415888] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:09:10.349 [2024-11-27 15:13:38.415916] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:09:10.349 [2024-11-27 15:13:38.416256] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:09:10.349 [2024-11-27 15:13:38.416418] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:09:10.349 [2024-11-27 15:13:38.416431] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:09:10.349 [2024-11-27 15:13:38.416663] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:10.349 BaseBdev3 00:09:10.349 15:13:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:10.349 15:13:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:10.349 15:13:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:10.349 15:13:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:10.349 15:13:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:10.349 15:13:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:10.349 15:13:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:10.349 15:13:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:10.349 15:13:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:10.349 15:13:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.349 15:13:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:10.349 15:13:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:10.349 15:13:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:10.349 15:13:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.349 [ 00:09:10.349 { 00:09:10.349 "name": "BaseBdev3", 00:09:10.349 "aliases": [ 00:09:10.349 "37899271-c110-46b3-86f6-28c4fa14ad26" 00:09:10.349 ], 00:09:10.349 "product_name": "Malloc disk", 00:09:10.349 "block_size": 512, 00:09:10.349 "num_blocks": 65536, 00:09:10.349 "uuid": "37899271-c110-46b3-86f6-28c4fa14ad26", 00:09:10.349 "assigned_rate_limits": { 00:09:10.349 "rw_ios_per_sec": 0, 00:09:10.349 "rw_mbytes_per_sec": 0, 00:09:10.349 "r_mbytes_per_sec": 0, 00:09:10.349 "w_mbytes_per_sec": 0 00:09:10.349 }, 00:09:10.349 "claimed": true, 00:09:10.349 "claim_type": "exclusive_write", 00:09:10.349 "zoned": false, 00:09:10.349 "supported_io_types": { 00:09:10.349 "read": true, 00:09:10.349 "write": true, 00:09:10.349 "unmap": true, 00:09:10.349 "flush": true, 00:09:10.349 "reset": true, 00:09:10.349 "nvme_admin": false, 00:09:10.349 "nvme_io": false, 00:09:10.349 "nvme_io_md": false, 00:09:10.349 "write_zeroes": true, 00:09:10.349 "zcopy": true, 00:09:10.349 "get_zone_info": false, 00:09:10.349 "zone_management": false, 00:09:10.349 "zone_append": false, 00:09:10.349 "compare": false, 00:09:10.349 "compare_and_write": false, 00:09:10.349 "abort": true, 00:09:10.349 "seek_hole": false, 00:09:10.349 "seek_data": false, 00:09:10.349 "copy": true, 00:09:10.349 "nvme_iov_md": false 00:09:10.349 }, 00:09:10.349 "memory_domains": [ 00:09:10.349 { 00:09:10.349 "dma_device_id": "system", 00:09:10.349 "dma_device_type": 1 00:09:10.349 }, 00:09:10.349 { 00:09:10.349 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:10.349 "dma_device_type": 2 00:09:10.349 } 00:09:10.349 ], 00:09:10.349 "driver_specific": {} 00:09:10.349 } 00:09:10.349 ] 00:09:10.349 15:13:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:10.349 15:13:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:10.349 15:13:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:10.609 15:13:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:10.609 15:13:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:09:10.609 15:13:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:10.609 15:13:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:10.609 15:13:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:10.609 15:13:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:10.609 15:13:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:10.609 15:13:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:10.609 15:13:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:10.609 15:13:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:10.609 15:13:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:10.609 15:13:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:10.609 15:13:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:10.609 15:13:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:10.609 15:13:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.609 15:13:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:10.609 15:13:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:10.609 "name": "Existed_Raid", 00:09:10.609 "uuid": "40cc40c0-44b1-484c-820e-cdbdc65b9767", 00:09:10.609 "strip_size_kb": 0, 00:09:10.609 "state": "online", 00:09:10.609 "raid_level": "raid1", 00:09:10.609 "superblock": false, 00:09:10.609 "num_base_bdevs": 3, 00:09:10.609 "num_base_bdevs_discovered": 3, 00:09:10.609 "num_base_bdevs_operational": 3, 00:09:10.609 "base_bdevs_list": [ 00:09:10.609 { 00:09:10.609 "name": "BaseBdev1", 00:09:10.609 "uuid": "492e46d2-9fca-44dc-8ca4-7664a9004c96", 00:09:10.609 "is_configured": true, 00:09:10.609 "data_offset": 0, 00:09:10.609 "data_size": 65536 00:09:10.609 }, 00:09:10.609 { 00:09:10.609 "name": "BaseBdev2", 00:09:10.609 "uuid": "0056499c-2eba-43bf-b9e2-348f9d8144af", 00:09:10.609 "is_configured": true, 00:09:10.609 "data_offset": 0, 00:09:10.609 "data_size": 65536 00:09:10.609 }, 00:09:10.609 { 00:09:10.609 "name": "BaseBdev3", 00:09:10.609 "uuid": "37899271-c110-46b3-86f6-28c4fa14ad26", 00:09:10.609 "is_configured": true, 00:09:10.609 "data_offset": 0, 00:09:10.609 "data_size": 65536 00:09:10.609 } 00:09:10.609 ] 00:09:10.609 }' 00:09:10.609 15:13:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:10.609 15:13:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.869 15:13:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:10.869 15:13:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:10.869 15:13:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:10.869 15:13:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:10.869 15:13:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:10.869 15:13:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:10.869 15:13:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:10.869 15:13:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:10.869 15:13:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:10.869 15:13:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.869 [2024-11-27 15:13:38.923501] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:10.869 15:13:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:10.869 15:13:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:10.869 "name": "Existed_Raid", 00:09:10.869 "aliases": [ 00:09:10.869 "40cc40c0-44b1-484c-820e-cdbdc65b9767" 00:09:10.869 ], 00:09:10.869 "product_name": "Raid Volume", 00:09:10.869 "block_size": 512, 00:09:10.869 "num_blocks": 65536, 00:09:10.869 "uuid": "40cc40c0-44b1-484c-820e-cdbdc65b9767", 00:09:10.869 "assigned_rate_limits": { 00:09:10.869 "rw_ios_per_sec": 0, 00:09:10.869 "rw_mbytes_per_sec": 0, 00:09:10.869 "r_mbytes_per_sec": 0, 00:09:10.869 "w_mbytes_per_sec": 0 00:09:10.869 }, 00:09:10.869 "claimed": false, 00:09:10.869 "zoned": false, 00:09:10.869 "supported_io_types": { 00:09:10.869 "read": true, 00:09:10.869 "write": true, 00:09:10.869 "unmap": false, 00:09:10.869 "flush": false, 00:09:10.869 "reset": true, 00:09:10.869 "nvme_admin": false, 00:09:10.869 "nvme_io": false, 00:09:10.869 "nvme_io_md": false, 00:09:10.869 "write_zeroes": true, 00:09:10.869 "zcopy": false, 00:09:10.869 "get_zone_info": false, 00:09:10.869 "zone_management": false, 00:09:10.869 "zone_append": false, 00:09:10.869 "compare": false, 00:09:10.869 "compare_and_write": false, 00:09:10.869 "abort": false, 00:09:10.869 "seek_hole": false, 00:09:10.869 "seek_data": false, 00:09:10.869 "copy": false, 00:09:10.869 "nvme_iov_md": false 00:09:10.869 }, 00:09:10.869 "memory_domains": [ 00:09:10.869 { 00:09:10.869 "dma_device_id": "system", 00:09:10.869 "dma_device_type": 1 00:09:10.869 }, 00:09:10.869 { 00:09:10.869 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:10.869 "dma_device_type": 2 00:09:10.869 }, 00:09:10.869 { 00:09:10.869 "dma_device_id": "system", 00:09:10.869 "dma_device_type": 1 00:09:10.869 }, 00:09:10.869 { 00:09:10.869 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:10.869 "dma_device_type": 2 00:09:10.869 }, 00:09:10.869 { 00:09:10.869 "dma_device_id": "system", 00:09:10.869 "dma_device_type": 1 00:09:10.869 }, 00:09:10.869 { 00:09:10.869 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:10.869 "dma_device_type": 2 00:09:10.869 } 00:09:10.869 ], 00:09:10.869 "driver_specific": { 00:09:10.869 "raid": { 00:09:10.869 "uuid": "40cc40c0-44b1-484c-820e-cdbdc65b9767", 00:09:10.869 "strip_size_kb": 0, 00:09:10.869 "state": "online", 00:09:10.869 "raid_level": "raid1", 00:09:10.869 "superblock": false, 00:09:10.869 "num_base_bdevs": 3, 00:09:10.869 "num_base_bdevs_discovered": 3, 00:09:10.869 "num_base_bdevs_operational": 3, 00:09:10.869 "base_bdevs_list": [ 00:09:10.869 { 00:09:10.869 "name": "BaseBdev1", 00:09:10.869 "uuid": "492e46d2-9fca-44dc-8ca4-7664a9004c96", 00:09:10.869 "is_configured": true, 00:09:10.869 "data_offset": 0, 00:09:10.869 "data_size": 65536 00:09:10.869 }, 00:09:10.869 { 00:09:10.869 "name": "BaseBdev2", 00:09:10.869 "uuid": "0056499c-2eba-43bf-b9e2-348f9d8144af", 00:09:10.869 "is_configured": true, 00:09:10.869 "data_offset": 0, 00:09:10.869 "data_size": 65536 00:09:10.869 }, 00:09:10.869 { 00:09:10.869 "name": "BaseBdev3", 00:09:10.869 "uuid": "37899271-c110-46b3-86f6-28c4fa14ad26", 00:09:10.869 "is_configured": true, 00:09:10.869 "data_offset": 0, 00:09:10.869 "data_size": 65536 00:09:10.869 } 00:09:10.869 ] 00:09:10.869 } 00:09:10.869 } 00:09:10.869 }' 00:09:10.869 15:13:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:11.130 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:11.130 BaseBdev2 00:09:11.130 BaseBdev3' 00:09:11.130 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:11.130 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:11.130 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:11.130 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:11.130 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:11.130 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.130 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.130 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.130 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:11.130 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:11.130 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:11.130 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:11.130 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.130 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.130 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:11.130 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.130 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:11.130 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:11.130 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:11.130 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:11.130 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:11.130 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.130 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.130 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.130 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:11.130 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:11.130 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:11.130 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.130 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.130 [2024-11-27 15:13:39.226691] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:11.389 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.389 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:11.389 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:09:11.389 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:11.389 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:11.389 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:09:11.389 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:09:11.389 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:11.389 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:11.389 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:11.389 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:11.389 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:11.389 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:11.389 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:11.389 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:11.389 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:11.389 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:11.389 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.389 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.389 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:11.389 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.389 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:11.389 "name": "Existed_Raid", 00:09:11.389 "uuid": "40cc40c0-44b1-484c-820e-cdbdc65b9767", 00:09:11.389 "strip_size_kb": 0, 00:09:11.389 "state": "online", 00:09:11.389 "raid_level": "raid1", 00:09:11.389 "superblock": false, 00:09:11.389 "num_base_bdevs": 3, 00:09:11.389 "num_base_bdevs_discovered": 2, 00:09:11.389 "num_base_bdevs_operational": 2, 00:09:11.389 "base_bdevs_list": [ 00:09:11.389 { 00:09:11.389 "name": null, 00:09:11.389 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:11.389 "is_configured": false, 00:09:11.389 "data_offset": 0, 00:09:11.389 "data_size": 65536 00:09:11.389 }, 00:09:11.389 { 00:09:11.389 "name": "BaseBdev2", 00:09:11.389 "uuid": "0056499c-2eba-43bf-b9e2-348f9d8144af", 00:09:11.389 "is_configured": true, 00:09:11.389 "data_offset": 0, 00:09:11.389 "data_size": 65536 00:09:11.389 }, 00:09:11.389 { 00:09:11.389 "name": "BaseBdev3", 00:09:11.389 "uuid": "37899271-c110-46b3-86f6-28c4fa14ad26", 00:09:11.389 "is_configured": true, 00:09:11.389 "data_offset": 0, 00:09:11.389 "data_size": 65536 00:09:11.389 } 00:09:11.389 ] 00:09:11.389 }' 00:09:11.389 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:11.389 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.647 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:11.647 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:11.647 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:11.647 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:11.647 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.647 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.647 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.647 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:11.647 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:11.647 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:11.647 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.647 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.647 [2024-11-27 15:13:39.745377] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:11.925 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.925 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:11.925 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:11.925 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:11.925 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:11.925 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.925 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.925 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.925 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:11.925 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:11.925 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:11.925 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.925 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.925 [2024-11-27 15:13:39.816917] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:11.925 [2024-11-27 15:13:39.817010] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:11.925 [2024-11-27 15:13:39.828676] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:11.925 [2024-11-27 15:13:39.828786] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:11.925 [2024-11-27 15:13:39.828833] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:09:11.925 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.925 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:11.925 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.926 BaseBdev2 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.926 [ 00:09:11.926 { 00:09:11.926 "name": "BaseBdev2", 00:09:11.926 "aliases": [ 00:09:11.926 "2a78981a-d0e3-4419-aeef-d784b8dcc737" 00:09:11.926 ], 00:09:11.926 "product_name": "Malloc disk", 00:09:11.926 "block_size": 512, 00:09:11.926 "num_blocks": 65536, 00:09:11.926 "uuid": "2a78981a-d0e3-4419-aeef-d784b8dcc737", 00:09:11.926 "assigned_rate_limits": { 00:09:11.926 "rw_ios_per_sec": 0, 00:09:11.926 "rw_mbytes_per_sec": 0, 00:09:11.926 "r_mbytes_per_sec": 0, 00:09:11.926 "w_mbytes_per_sec": 0 00:09:11.926 }, 00:09:11.926 "claimed": false, 00:09:11.926 "zoned": false, 00:09:11.926 "supported_io_types": { 00:09:11.926 "read": true, 00:09:11.926 "write": true, 00:09:11.926 "unmap": true, 00:09:11.926 "flush": true, 00:09:11.926 "reset": true, 00:09:11.926 "nvme_admin": false, 00:09:11.926 "nvme_io": false, 00:09:11.926 "nvme_io_md": false, 00:09:11.926 "write_zeroes": true, 00:09:11.926 "zcopy": true, 00:09:11.926 "get_zone_info": false, 00:09:11.926 "zone_management": false, 00:09:11.926 "zone_append": false, 00:09:11.926 "compare": false, 00:09:11.926 "compare_and_write": false, 00:09:11.926 "abort": true, 00:09:11.926 "seek_hole": false, 00:09:11.926 "seek_data": false, 00:09:11.926 "copy": true, 00:09:11.926 "nvme_iov_md": false 00:09:11.926 }, 00:09:11.926 "memory_domains": [ 00:09:11.926 { 00:09:11.926 "dma_device_id": "system", 00:09:11.926 "dma_device_type": 1 00:09:11.926 }, 00:09:11.926 { 00:09:11.926 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:11.926 "dma_device_type": 2 00:09:11.926 } 00:09:11.926 ], 00:09:11.926 "driver_specific": {} 00:09:11.926 } 00:09:11.926 ] 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.926 BaseBdev3 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.926 [ 00:09:11.926 { 00:09:11.926 "name": "BaseBdev3", 00:09:11.926 "aliases": [ 00:09:11.926 "85852961-2c1e-4b00-ae31-116ac882afa6" 00:09:11.926 ], 00:09:11.926 "product_name": "Malloc disk", 00:09:11.926 "block_size": 512, 00:09:11.926 "num_blocks": 65536, 00:09:11.926 "uuid": "85852961-2c1e-4b00-ae31-116ac882afa6", 00:09:11.926 "assigned_rate_limits": { 00:09:11.926 "rw_ios_per_sec": 0, 00:09:11.926 "rw_mbytes_per_sec": 0, 00:09:11.926 "r_mbytes_per_sec": 0, 00:09:11.926 "w_mbytes_per_sec": 0 00:09:11.926 }, 00:09:11.926 "claimed": false, 00:09:11.926 "zoned": false, 00:09:11.926 "supported_io_types": { 00:09:11.926 "read": true, 00:09:11.926 "write": true, 00:09:11.926 "unmap": true, 00:09:11.926 "flush": true, 00:09:11.926 "reset": true, 00:09:11.926 "nvme_admin": false, 00:09:11.926 "nvme_io": false, 00:09:11.926 "nvme_io_md": false, 00:09:11.926 "write_zeroes": true, 00:09:11.926 "zcopy": true, 00:09:11.926 "get_zone_info": false, 00:09:11.926 "zone_management": false, 00:09:11.926 "zone_append": false, 00:09:11.926 "compare": false, 00:09:11.926 "compare_and_write": false, 00:09:11.926 "abort": true, 00:09:11.926 "seek_hole": false, 00:09:11.926 "seek_data": false, 00:09:11.926 "copy": true, 00:09:11.926 "nvme_iov_md": false 00:09:11.926 }, 00:09:11.926 "memory_domains": [ 00:09:11.926 { 00:09:11.926 "dma_device_id": "system", 00:09:11.926 "dma_device_type": 1 00:09:11.926 }, 00:09:11.926 { 00:09:11.926 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:11.926 "dma_device_type": 2 00:09:11.926 } 00:09:11.926 ], 00:09:11.926 "driver_specific": {} 00:09:11.926 } 00:09:11.926 ] 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.926 [2024-11-27 15:13:39.985321] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:11.926 [2024-11-27 15:13:39.985391] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:11.926 [2024-11-27 15:13:39.985425] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:11.926 [2024-11-27 15:13:39.987367] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:11.926 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:11.927 15:13:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:11.927 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.927 15:13:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.927 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:12.218 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:12.218 "name": "Existed_Raid", 00:09:12.218 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:12.218 "strip_size_kb": 0, 00:09:12.218 "state": "configuring", 00:09:12.218 "raid_level": "raid1", 00:09:12.218 "superblock": false, 00:09:12.218 "num_base_bdevs": 3, 00:09:12.218 "num_base_bdevs_discovered": 2, 00:09:12.218 "num_base_bdevs_operational": 3, 00:09:12.218 "base_bdevs_list": [ 00:09:12.218 { 00:09:12.218 "name": "BaseBdev1", 00:09:12.218 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:12.218 "is_configured": false, 00:09:12.218 "data_offset": 0, 00:09:12.218 "data_size": 0 00:09:12.218 }, 00:09:12.218 { 00:09:12.218 "name": "BaseBdev2", 00:09:12.218 "uuid": "2a78981a-d0e3-4419-aeef-d784b8dcc737", 00:09:12.218 "is_configured": true, 00:09:12.218 "data_offset": 0, 00:09:12.218 "data_size": 65536 00:09:12.218 }, 00:09:12.218 { 00:09:12.218 "name": "BaseBdev3", 00:09:12.218 "uuid": "85852961-2c1e-4b00-ae31-116ac882afa6", 00:09:12.218 "is_configured": true, 00:09:12.218 "data_offset": 0, 00:09:12.218 "data_size": 65536 00:09:12.218 } 00:09:12.218 ] 00:09:12.218 }' 00:09:12.218 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:12.218 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.477 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:12.477 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:12.477 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.477 [2024-11-27 15:13:40.440613] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:12.477 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:12.477 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:12.477 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:12.477 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:12.477 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:12.477 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:12.477 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:12.477 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:12.477 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:12.477 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:12.477 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:12.477 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:12.477 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:12.477 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:12.477 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.477 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:12.477 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:12.477 "name": "Existed_Raid", 00:09:12.477 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:12.477 "strip_size_kb": 0, 00:09:12.477 "state": "configuring", 00:09:12.477 "raid_level": "raid1", 00:09:12.477 "superblock": false, 00:09:12.477 "num_base_bdevs": 3, 00:09:12.477 "num_base_bdevs_discovered": 1, 00:09:12.477 "num_base_bdevs_operational": 3, 00:09:12.477 "base_bdevs_list": [ 00:09:12.477 { 00:09:12.477 "name": "BaseBdev1", 00:09:12.477 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:12.477 "is_configured": false, 00:09:12.477 "data_offset": 0, 00:09:12.477 "data_size": 0 00:09:12.477 }, 00:09:12.477 { 00:09:12.477 "name": null, 00:09:12.477 "uuid": "2a78981a-d0e3-4419-aeef-d784b8dcc737", 00:09:12.477 "is_configured": false, 00:09:12.477 "data_offset": 0, 00:09:12.477 "data_size": 65536 00:09:12.477 }, 00:09:12.477 { 00:09:12.477 "name": "BaseBdev3", 00:09:12.477 "uuid": "85852961-2c1e-4b00-ae31-116ac882afa6", 00:09:12.477 "is_configured": true, 00:09:12.477 "data_offset": 0, 00:09:12.477 "data_size": 65536 00:09:12.477 } 00:09:12.477 ] 00:09:12.477 }' 00:09:12.477 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:12.477 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.737 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:12.737 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:12.737 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:12.737 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.997 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:12.997 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:12.997 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:12.997 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:12.997 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.997 [2024-11-27 15:13:40.875084] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:12.997 BaseBdev1 00:09:12.997 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:12.997 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:12.997 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:12.997 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:12.997 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:12.997 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:12.997 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:12.997 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:12.997 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:12.997 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.997 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:12.997 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:12.997 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:12.997 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.997 [ 00:09:12.997 { 00:09:12.997 "name": "BaseBdev1", 00:09:12.997 "aliases": [ 00:09:12.997 "38b1b4b9-bb79-4bfc-b282-2799042a5598" 00:09:12.997 ], 00:09:12.997 "product_name": "Malloc disk", 00:09:12.997 "block_size": 512, 00:09:12.997 "num_blocks": 65536, 00:09:12.997 "uuid": "38b1b4b9-bb79-4bfc-b282-2799042a5598", 00:09:12.997 "assigned_rate_limits": { 00:09:12.997 "rw_ios_per_sec": 0, 00:09:12.997 "rw_mbytes_per_sec": 0, 00:09:12.997 "r_mbytes_per_sec": 0, 00:09:12.997 "w_mbytes_per_sec": 0 00:09:12.997 }, 00:09:12.997 "claimed": true, 00:09:12.997 "claim_type": "exclusive_write", 00:09:12.997 "zoned": false, 00:09:12.997 "supported_io_types": { 00:09:12.997 "read": true, 00:09:12.997 "write": true, 00:09:12.997 "unmap": true, 00:09:12.997 "flush": true, 00:09:12.997 "reset": true, 00:09:12.997 "nvme_admin": false, 00:09:12.997 "nvme_io": false, 00:09:12.997 "nvme_io_md": false, 00:09:12.998 "write_zeroes": true, 00:09:12.998 "zcopy": true, 00:09:12.998 "get_zone_info": false, 00:09:12.998 "zone_management": false, 00:09:12.998 "zone_append": false, 00:09:12.998 "compare": false, 00:09:12.998 "compare_and_write": false, 00:09:12.998 "abort": true, 00:09:12.998 "seek_hole": false, 00:09:12.998 "seek_data": false, 00:09:12.998 "copy": true, 00:09:12.998 "nvme_iov_md": false 00:09:12.998 }, 00:09:12.998 "memory_domains": [ 00:09:12.998 { 00:09:12.998 "dma_device_id": "system", 00:09:12.998 "dma_device_type": 1 00:09:12.998 }, 00:09:12.998 { 00:09:12.998 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:12.998 "dma_device_type": 2 00:09:12.998 } 00:09:12.998 ], 00:09:12.998 "driver_specific": {} 00:09:12.998 } 00:09:12.998 ] 00:09:12.998 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:12.998 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:12.998 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:12.998 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:12.998 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:12.998 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:12.998 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:12.998 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:12.998 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:12.998 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:12.998 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:12.998 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:12.998 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:12.998 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:12.998 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:12.998 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.998 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:12.998 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:12.998 "name": "Existed_Raid", 00:09:12.998 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:12.998 "strip_size_kb": 0, 00:09:12.998 "state": "configuring", 00:09:12.998 "raid_level": "raid1", 00:09:12.998 "superblock": false, 00:09:12.998 "num_base_bdevs": 3, 00:09:12.998 "num_base_bdevs_discovered": 2, 00:09:12.998 "num_base_bdevs_operational": 3, 00:09:12.998 "base_bdevs_list": [ 00:09:12.998 { 00:09:12.998 "name": "BaseBdev1", 00:09:12.998 "uuid": "38b1b4b9-bb79-4bfc-b282-2799042a5598", 00:09:12.998 "is_configured": true, 00:09:12.998 "data_offset": 0, 00:09:12.998 "data_size": 65536 00:09:12.998 }, 00:09:12.998 { 00:09:12.998 "name": null, 00:09:12.998 "uuid": "2a78981a-d0e3-4419-aeef-d784b8dcc737", 00:09:12.998 "is_configured": false, 00:09:12.998 "data_offset": 0, 00:09:12.998 "data_size": 65536 00:09:12.998 }, 00:09:12.998 { 00:09:12.998 "name": "BaseBdev3", 00:09:12.998 "uuid": "85852961-2c1e-4b00-ae31-116ac882afa6", 00:09:12.998 "is_configured": true, 00:09:12.998 "data_offset": 0, 00:09:12.998 "data_size": 65536 00:09:12.998 } 00:09:12.998 ] 00:09:12.998 }' 00:09:12.998 15:13:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:12.998 15:13:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.258 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:13.258 15:13:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:13.258 15:13:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.258 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:13.516 15:13:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:13.516 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:13.517 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:13.517 15:13:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:13.517 15:13:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.517 [2024-11-27 15:13:41.410234] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:13.517 15:13:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:13.517 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:13.517 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:13.517 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:13.517 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:13.517 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:13.517 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:13.517 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:13.517 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:13.517 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:13.517 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:13.517 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:13.517 15:13:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:13.517 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:13.517 15:13:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.517 15:13:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:13.517 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:13.517 "name": "Existed_Raid", 00:09:13.517 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:13.517 "strip_size_kb": 0, 00:09:13.517 "state": "configuring", 00:09:13.517 "raid_level": "raid1", 00:09:13.517 "superblock": false, 00:09:13.517 "num_base_bdevs": 3, 00:09:13.517 "num_base_bdevs_discovered": 1, 00:09:13.517 "num_base_bdevs_operational": 3, 00:09:13.517 "base_bdevs_list": [ 00:09:13.517 { 00:09:13.517 "name": "BaseBdev1", 00:09:13.517 "uuid": "38b1b4b9-bb79-4bfc-b282-2799042a5598", 00:09:13.517 "is_configured": true, 00:09:13.517 "data_offset": 0, 00:09:13.517 "data_size": 65536 00:09:13.517 }, 00:09:13.517 { 00:09:13.517 "name": null, 00:09:13.517 "uuid": "2a78981a-d0e3-4419-aeef-d784b8dcc737", 00:09:13.517 "is_configured": false, 00:09:13.517 "data_offset": 0, 00:09:13.517 "data_size": 65536 00:09:13.517 }, 00:09:13.517 { 00:09:13.517 "name": null, 00:09:13.517 "uuid": "85852961-2c1e-4b00-ae31-116ac882afa6", 00:09:13.517 "is_configured": false, 00:09:13.517 "data_offset": 0, 00:09:13.517 "data_size": 65536 00:09:13.517 } 00:09:13.517 ] 00:09:13.517 }' 00:09:13.517 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:13.517 15:13:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.086 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:14.086 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:14.086 15:13:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.086 15:13:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.086 15:13:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.086 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:14.086 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:14.086 15:13:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.086 15:13:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.086 [2024-11-27 15:13:41.929382] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:14.086 15:13:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.086 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:14.086 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:14.086 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:14.086 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:14.086 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:14.086 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:14.086 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:14.086 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:14.086 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:14.086 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:14.086 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:14.086 15:13:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.086 15:13:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.086 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:14.086 15:13:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.086 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:14.086 "name": "Existed_Raid", 00:09:14.086 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:14.086 "strip_size_kb": 0, 00:09:14.086 "state": "configuring", 00:09:14.086 "raid_level": "raid1", 00:09:14.086 "superblock": false, 00:09:14.086 "num_base_bdevs": 3, 00:09:14.086 "num_base_bdevs_discovered": 2, 00:09:14.086 "num_base_bdevs_operational": 3, 00:09:14.086 "base_bdevs_list": [ 00:09:14.086 { 00:09:14.086 "name": "BaseBdev1", 00:09:14.086 "uuid": "38b1b4b9-bb79-4bfc-b282-2799042a5598", 00:09:14.086 "is_configured": true, 00:09:14.086 "data_offset": 0, 00:09:14.086 "data_size": 65536 00:09:14.086 }, 00:09:14.086 { 00:09:14.086 "name": null, 00:09:14.086 "uuid": "2a78981a-d0e3-4419-aeef-d784b8dcc737", 00:09:14.086 "is_configured": false, 00:09:14.086 "data_offset": 0, 00:09:14.086 "data_size": 65536 00:09:14.086 }, 00:09:14.086 { 00:09:14.086 "name": "BaseBdev3", 00:09:14.086 "uuid": "85852961-2c1e-4b00-ae31-116ac882afa6", 00:09:14.086 "is_configured": true, 00:09:14.086 "data_offset": 0, 00:09:14.086 "data_size": 65536 00:09:14.086 } 00:09:14.086 ] 00:09:14.086 }' 00:09:14.086 15:13:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:14.086 15:13:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.346 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:14.346 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:14.346 15:13:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.346 15:13:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.346 15:13:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.346 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:14.346 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:14.346 15:13:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.346 15:13:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.346 [2024-11-27 15:13:42.412700] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:14.346 15:13:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.346 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:14.346 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:14.346 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:14.346 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:14.346 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:14.346 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:14.346 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:14.346 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:14.346 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:14.346 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:14.346 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:14.346 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:14.346 15:13:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.346 15:13:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.346 15:13:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.606 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:14.606 "name": "Existed_Raid", 00:09:14.606 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:14.606 "strip_size_kb": 0, 00:09:14.606 "state": "configuring", 00:09:14.606 "raid_level": "raid1", 00:09:14.606 "superblock": false, 00:09:14.606 "num_base_bdevs": 3, 00:09:14.606 "num_base_bdevs_discovered": 1, 00:09:14.606 "num_base_bdevs_operational": 3, 00:09:14.606 "base_bdevs_list": [ 00:09:14.606 { 00:09:14.606 "name": null, 00:09:14.606 "uuid": "38b1b4b9-bb79-4bfc-b282-2799042a5598", 00:09:14.606 "is_configured": false, 00:09:14.606 "data_offset": 0, 00:09:14.606 "data_size": 65536 00:09:14.606 }, 00:09:14.606 { 00:09:14.606 "name": null, 00:09:14.606 "uuid": "2a78981a-d0e3-4419-aeef-d784b8dcc737", 00:09:14.606 "is_configured": false, 00:09:14.606 "data_offset": 0, 00:09:14.606 "data_size": 65536 00:09:14.606 }, 00:09:14.606 { 00:09:14.606 "name": "BaseBdev3", 00:09:14.606 "uuid": "85852961-2c1e-4b00-ae31-116ac882afa6", 00:09:14.606 "is_configured": true, 00:09:14.606 "data_offset": 0, 00:09:14.606 "data_size": 65536 00:09:14.606 } 00:09:14.606 ] 00:09:14.606 }' 00:09:14.606 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:14.606 15:13:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.867 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:14.867 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:14.867 15:13:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.867 15:13:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.867 15:13:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.867 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:14.867 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:14.867 15:13:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.867 15:13:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.867 [2024-11-27 15:13:42.950377] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:14.867 15:13:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.867 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:14.867 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:14.867 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:14.867 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:14.867 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:14.867 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:14.867 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:14.867 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:14.867 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:14.867 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:14.867 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:14.867 15:13:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.867 15:13:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.867 15:13:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:14.867 15:13:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:15.127 15:13:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:15.127 "name": "Existed_Raid", 00:09:15.127 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:15.127 "strip_size_kb": 0, 00:09:15.127 "state": "configuring", 00:09:15.127 "raid_level": "raid1", 00:09:15.127 "superblock": false, 00:09:15.127 "num_base_bdevs": 3, 00:09:15.127 "num_base_bdevs_discovered": 2, 00:09:15.127 "num_base_bdevs_operational": 3, 00:09:15.127 "base_bdevs_list": [ 00:09:15.127 { 00:09:15.127 "name": null, 00:09:15.127 "uuid": "38b1b4b9-bb79-4bfc-b282-2799042a5598", 00:09:15.127 "is_configured": false, 00:09:15.127 "data_offset": 0, 00:09:15.127 "data_size": 65536 00:09:15.127 }, 00:09:15.127 { 00:09:15.127 "name": "BaseBdev2", 00:09:15.127 "uuid": "2a78981a-d0e3-4419-aeef-d784b8dcc737", 00:09:15.127 "is_configured": true, 00:09:15.127 "data_offset": 0, 00:09:15.127 "data_size": 65536 00:09:15.127 }, 00:09:15.127 { 00:09:15.127 "name": "BaseBdev3", 00:09:15.127 "uuid": "85852961-2c1e-4b00-ae31-116ac882afa6", 00:09:15.127 "is_configured": true, 00:09:15.127 "data_offset": 0, 00:09:15.127 "data_size": 65536 00:09:15.127 } 00:09:15.127 ] 00:09:15.127 }' 00:09:15.127 15:13:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:15.127 15:13:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.388 15:13:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:15.388 15:13:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:15.388 15:13:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:15.388 15:13:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.388 15:13:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:15.388 15:13:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:15.388 15:13:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:15.388 15:13:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:15.388 15:13:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:15.388 15:13:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.388 15:13:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:15.388 15:13:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 38b1b4b9-bb79-4bfc-b282-2799042a5598 00:09:15.388 15:13:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:15.388 15:13:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.388 [2024-11-27 15:13:43.468636] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:15.388 [2024-11-27 15:13:43.468695] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:09:15.388 [2024-11-27 15:13:43.468702] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:09:15.388 [2024-11-27 15:13:43.468995] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:09:15.388 [2024-11-27 15:13:43.469132] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:09:15.388 [2024-11-27 15:13:43.469154] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006d00 00:09:15.388 [2024-11-27 15:13:43.469352] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:15.388 NewBaseBdev 00:09:15.388 15:13:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:15.388 15:13:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:15.388 15:13:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:09:15.388 15:13:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:15.388 15:13:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:15.388 15:13:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:15.388 15:13:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:15.388 15:13:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:15.388 15:13:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:15.388 15:13:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.388 15:13:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:15.388 15:13:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:15.388 15:13:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:15.388 15:13:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.647 [ 00:09:15.648 { 00:09:15.648 "name": "NewBaseBdev", 00:09:15.648 "aliases": [ 00:09:15.648 "38b1b4b9-bb79-4bfc-b282-2799042a5598" 00:09:15.648 ], 00:09:15.648 "product_name": "Malloc disk", 00:09:15.648 "block_size": 512, 00:09:15.648 "num_blocks": 65536, 00:09:15.648 "uuid": "38b1b4b9-bb79-4bfc-b282-2799042a5598", 00:09:15.648 "assigned_rate_limits": { 00:09:15.648 "rw_ios_per_sec": 0, 00:09:15.648 "rw_mbytes_per_sec": 0, 00:09:15.648 "r_mbytes_per_sec": 0, 00:09:15.648 "w_mbytes_per_sec": 0 00:09:15.648 }, 00:09:15.648 "claimed": true, 00:09:15.648 "claim_type": "exclusive_write", 00:09:15.648 "zoned": false, 00:09:15.648 "supported_io_types": { 00:09:15.648 "read": true, 00:09:15.648 "write": true, 00:09:15.648 "unmap": true, 00:09:15.648 "flush": true, 00:09:15.648 "reset": true, 00:09:15.648 "nvme_admin": false, 00:09:15.648 "nvme_io": false, 00:09:15.648 "nvme_io_md": false, 00:09:15.648 "write_zeroes": true, 00:09:15.648 "zcopy": true, 00:09:15.648 "get_zone_info": false, 00:09:15.648 "zone_management": false, 00:09:15.648 "zone_append": false, 00:09:15.648 "compare": false, 00:09:15.648 "compare_and_write": false, 00:09:15.648 "abort": true, 00:09:15.648 "seek_hole": false, 00:09:15.648 "seek_data": false, 00:09:15.648 "copy": true, 00:09:15.648 "nvme_iov_md": false 00:09:15.648 }, 00:09:15.648 "memory_domains": [ 00:09:15.648 { 00:09:15.648 "dma_device_id": "system", 00:09:15.648 "dma_device_type": 1 00:09:15.648 }, 00:09:15.648 { 00:09:15.648 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:15.648 "dma_device_type": 2 00:09:15.648 } 00:09:15.648 ], 00:09:15.648 "driver_specific": {} 00:09:15.648 } 00:09:15.648 ] 00:09:15.648 15:13:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:15.648 15:13:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:15.648 15:13:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:09:15.648 15:13:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:15.648 15:13:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:15.648 15:13:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:15.648 15:13:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:15.648 15:13:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:15.648 15:13:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:15.648 15:13:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:15.648 15:13:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:15.648 15:13:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:15.648 15:13:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:15.648 15:13:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:15.648 15:13:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:15.648 15:13:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.648 15:13:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:15.648 15:13:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:15.648 "name": "Existed_Raid", 00:09:15.648 "uuid": "4795981a-5a90-4bee-add5-adf3235ed000", 00:09:15.648 "strip_size_kb": 0, 00:09:15.648 "state": "online", 00:09:15.648 "raid_level": "raid1", 00:09:15.648 "superblock": false, 00:09:15.648 "num_base_bdevs": 3, 00:09:15.648 "num_base_bdevs_discovered": 3, 00:09:15.648 "num_base_bdevs_operational": 3, 00:09:15.648 "base_bdevs_list": [ 00:09:15.648 { 00:09:15.648 "name": "NewBaseBdev", 00:09:15.648 "uuid": "38b1b4b9-bb79-4bfc-b282-2799042a5598", 00:09:15.648 "is_configured": true, 00:09:15.648 "data_offset": 0, 00:09:15.648 "data_size": 65536 00:09:15.648 }, 00:09:15.648 { 00:09:15.648 "name": "BaseBdev2", 00:09:15.648 "uuid": "2a78981a-d0e3-4419-aeef-d784b8dcc737", 00:09:15.648 "is_configured": true, 00:09:15.648 "data_offset": 0, 00:09:15.648 "data_size": 65536 00:09:15.648 }, 00:09:15.648 { 00:09:15.648 "name": "BaseBdev3", 00:09:15.648 "uuid": "85852961-2c1e-4b00-ae31-116ac882afa6", 00:09:15.648 "is_configured": true, 00:09:15.648 "data_offset": 0, 00:09:15.648 "data_size": 65536 00:09:15.648 } 00:09:15.648 ] 00:09:15.648 }' 00:09:15.648 15:13:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:15.648 15:13:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.939 15:13:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:15.939 15:13:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:15.939 15:13:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:15.939 15:13:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:15.939 15:13:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:15.939 15:13:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:15.939 15:13:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:15.939 15:13:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:15.939 15:13:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.939 15:13:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:15.939 [2024-11-27 15:13:43.984143] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:15.939 15:13:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:15.939 15:13:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:15.939 "name": "Existed_Raid", 00:09:15.939 "aliases": [ 00:09:15.939 "4795981a-5a90-4bee-add5-adf3235ed000" 00:09:15.939 ], 00:09:15.939 "product_name": "Raid Volume", 00:09:15.939 "block_size": 512, 00:09:15.939 "num_blocks": 65536, 00:09:15.939 "uuid": "4795981a-5a90-4bee-add5-adf3235ed000", 00:09:15.939 "assigned_rate_limits": { 00:09:15.939 "rw_ios_per_sec": 0, 00:09:15.939 "rw_mbytes_per_sec": 0, 00:09:15.939 "r_mbytes_per_sec": 0, 00:09:15.939 "w_mbytes_per_sec": 0 00:09:15.939 }, 00:09:15.939 "claimed": false, 00:09:15.939 "zoned": false, 00:09:15.939 "supported_io_types": { 00:09:15.939 "read": true, 00:09:15.939 "write": true, 00:09:15.939 "unmap": false, 00:09:15.939 "flush": false, 00:09:15.939 "reset": true, 00:09:15.939 "nvme_admin": false, 00:09:15.939 "nvme_io": false, 00:09:15.939 "nvme_io_md": false, 00:09:15.939 "write_zeroes": true, 00:09:15.939 "zcopy": false, 00:09:15.939 "get_zone_info": false, 00:09:15.939 "zone_management": false, 00:09:15.939 "zone_append": false, 00:09:15.939 "compare": false, 00:09:15.939 "compare_and_write": false, 00:09:15.939 "abort": false, 00:09:15.939 "seek_hole": false, 00:09:15.939 "seek_data": false, 00:09:15.939 "copy": false, 00:09:15.939 "nvme_iov_md": false 00:09:15.939 }, 00:09:15.939 "memory_domains": [ 00:09:15.939 { 00:09:15.939 "dma_device_id": "system", 00:09:15.939 "dma_device_type": 1 00:09:15.939 }, 00:09:15.939 { 00:09:15.939 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:15.939 "dma_device_type": 2 00:09:15.939 }, 00:09:15.939 { 00:09:15.939 "dma_device_id": "system", 00:09:15.939 "dma_device_type": 1 00:09:15.939 }, 00:09:15.939 { 00:09:15.939 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:15.939 "dma_device_type": 2 00:09:15.939 }, 00:09:15.939 { 00:09:15.939 "dma_device_id": "system", 00:09:15.939 "dma_device_type": 1 00:09:15.939 }, 00:09:15.939 { 00:09:15.939 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:15.939 "dma_device_type": 2 00:09:15.939 } 00:09:15.939 ], 00:09:15.939 "driver_specific": { 00:09:15.939 "raid": { 00:09:15.939 "uuid": "4795981a-5a90-4bee-add5-adf3235ed000", 00:09:15.939 "strip_size_kb": 0, 00:09:15.939 "state": "online", 00:09:15.939 "raid_level": "raid1", 00:09:15.939 "superblock": false, 00:09:15.939 "num_base_bdevs": 3, 00:09:15.939 "num_base_bdevs_discovered": 3, 00:09:15.939 "num_base_bdevs_operational": 3, 00:09:15.939 "base_bdevs_list": [ 00:09:15.939 { 00:09:15.939 "name": "NewBaseBdev", 00:09:15.939 "uuid": "38b1b4b9-bb79-4bfc-b282-2799042a5598", 00:09:15.939 "is_configured": true, 00:09:15.939 "data_offset": 0, 00:09:15.939 "data_size": 65536 00:09:15.939 }, 00:09:15.939 { 00:09:15.939 "name": "BaseBdev2", 00:09:15.939 "uuid": "2a78981a-d0e3-4419-aeef-d784b8dcc737", 00:09:15.939 "is_configured": true, 00:09:15.939 "data_offset": 0, 00:09:15.939 "data_size": 65536 00:09:15.939 }, 00:09:15.939 { 00:09:15.939 "name": "BaseBdev3", 00:09:15.939 "uuid": "85852961-2c1e-4b00-ae31-116ac882afa6", 00:09:15.939 "is_configured": true, 00:09:15.939 "data_offset": 0, 00:09:15.939 "data_size": 65536 00:09:15.939 } 00:09:15.939 ] 00:09:15.939 } 00:09:15.939 } 00:09:15.939 }' 00:09:15.939 15:13:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:16.200 BaseBdev2 00:09:16.200 BaseBdev3' 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.200 [2024-11-27 15:13:44.263515] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:16.200 [2024-11-27 15:13:44.263554] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:16.200 [2024-11-27 15:13:44.263645] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:16.200 [2024-11-27 15:13:44.263911] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:16.200 [2024-11-27 15:13:44.263933] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name Existed_Raid, state offline 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 78619 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 78619 ']' 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 78619 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:16.200 15:13:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 78619 00:09:16.460 15:13:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:16.460 15:13:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:16.460 killing process with pid 78619 00:09:16.460 15:13:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 78619' 00:09:16.460 15:13:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 78619 00:09:16.460 [2024-11-27 15:13:44.314516] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:16.460 15:13:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 78619 00:09:16.460 [2024-11-27 15:13:44.344842] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:16.460 15:13:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:09:16.460 00:09:16.460 real 0m8.983s 00:09:16.460 user 0m15.339s 00:09:16.460 sys 0m1.906s 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.721 ************************************ 00:09:16.721 END TEST raid_state_function_test 00:09:16.721 ************************************ 00:09:16.721 15:13:44 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 3 true 00:09:16.721 15:13:44 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:16.721 15:13:44 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:16.721 15:13:44 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:16.721 ************************************ 00:09:16.721 START TEST raid_state_function_test_sb 00:09:16.721 ************************************ 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 3 true 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=79224 00:09:16.721 Process raid pid: 79224 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 79224' 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 79224 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 79224 ']' 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:16.721 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:16.721 15:13:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:16.721 [2024-11-27 15:13:44.730684] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:09:16.721 [2024-11-27 15:13:44.730811] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:16.981 [2024-11-27 15:13:44.902049] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:16.981 [2024-11-27 15:13:44.927780] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:16.981 [2024-11-27 15:13:44.970878] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:16.981 [2024-11-27 15:13:44.970936] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:17.551 15:13:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:17.551 15:13:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:09:17.551 15:13:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:17.551 15:13:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.551 15:13:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:17.551 [2024-11-27 15:13:45.582307] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:17.551 [2024-11-27 15:13:45.582382] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:17.551 [2024-11-27 15:13:45.582392] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:17.551 [2024-11-27 15:13:45.582401] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:17.551 [2024-11-27 15:13:45.582407] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:17.551 [2024-11-27 15:13:45.582417] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:17.551 15:13:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:17.551 15:13:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:17.551 15:13:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:17.551 15:13:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:17.551 15:13:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:17.551 15:13:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:17.551 15:13:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:17.551 15:13:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:17.551 15:13:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:17.551 15:13:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:17.551 15:13:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:17.551 15:13:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:17.551 15:13:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:17.551 15:13:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.551 15:13:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:17.551 15:13:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:17.551 15:13:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:17.551 "name": "Existed_Raid", 00:09:17.551 "uuid": "819f5d9a-0b5f-402f-8b4d-45526f3e0df2", 00:09:17.551 "strip_size_kb": 0, 00:09:17.551 "state": "configuring", 00:09:17.551 "raid_level": "raid1", 00:09:17.551 "superblock": true, 00:09:17.551 "num_base_bdevs": 3, 00:09:17.551 "num_base_bdevs_discovered": 0, 00:09:17.551 "num_base_bdevs_operational": 3, 00:09:17.551 "base_bdevs_list": [ 00:09:17.551 { 00:09:17.551 "name": "BaseBdev1", 00:09:17.551 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:17.551 "is_configured": false, 00:09:17.551 "data_offset": 0, 00:09:17.551 "data_size": 0 00:09:17.551 }, 00:09:17.551 { 00:09:17.551 "name": "BaseBdev2", 00:09:17.551 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:17.551 "is_configured": false, 00:09:17.551 "data_offset": 0, 00:09:17.551 "data_size": 0 00:09:17.551 }, 00:09:17.551 { 00:09:17.551 "name": "BaseBdev3", 00:09:17.551 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:17.551 "is_configured": false, 00:09:17.551 "data_offset": 0, 00:09:17.551 "data_size": 0 00:09:17.551 } 00:09:17.551 ] 00:09:17.551 }' 00:09:17.551 15:13:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:17.551 15:13:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.121 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:18.121 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.121 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.121 [2024-11-27 15:13:46.005483] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:18.121 [2024-11-27 15:13:46.005530] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:09:18.121 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.121 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:18.121 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.121 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.121 [2024-11-27 15:13:46.013453] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:18.121 [2024-11-27 15:13:46.013490] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:18.121 [2024-11-27 15:13:46.013499] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:18.121 [2024-11-27 15:13:46.013508] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:18.121 [2024-11-27 15:13:46.013514] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:18.121 [2024-11-27 15:13:46.013523] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:18.121 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.121 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:18.121 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.121 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.121 [2024-11-27 15:13:46.030363] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:18.121 BaseBdev1 00:09:18.121 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.121 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:18.121 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:18.121 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:18.121 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:18.121 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:18.121 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:18.121 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:18.121 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.121 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.121 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.121 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:18.121 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.121 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.121 [ 00:09:18.121 { 00:09:18.121 "name": "BaseBdev1", 00:09:18.121 "aliases": [ 00:09:18.122 "f17f1bfa-8f3d-46a6-8c64-428af4dc5e45" 00:09:18.122 ], 00:09:18.122 "product_name": "Malloc disk", 00:09:18.122 "block_size": 512, 00:09:18.122 "num_blocks": 65536, 00:09:18.122 "uuid": "f17f1bfa-8f3d-46a6-8c64-428af4dc5e45", 00:09:18.122 "assigned_rate_limits": { 00:09:18.122 "rw_ios_per_sec": 0, 00:09:18.122 "rw_mbytes_per_sec": 0, 00:09:18.122 "r_mbytes_per_sec": 0, 00:09:18.122 "w_mbytes_per_sec": 0 00:09:18.122 }, 00:09:18.122 "claimed": true, 00:09:18.122 "claim_type": "exclusive_write", 00:09:18.122 "zoned": false, 00:09:18.122 "supported_io_types": { 00:09:18.122 "read": true, 00:09:18.122 "write": true, 00:09:18.122 "unmap": true, 00:09:18.122 "flush": true, 00:09:18.122 "reset": true, 00:09:18.122 "nvme_admin": false, 00:09:18.122 "nvme_io": false, 00:09:18.122 "nvme_io_md": false, 00:09:18.122 "write_zeroes": true, 00:09:18.122 "zcopy": true, 00:09:18.122 "get_zone_info": false, 00:09:18.122 "zone_management": false, 00:09:18.122 "zone_append": false, 00:09:18.122 "compare": false, 00:09:18.122 "compare_and_write": false, 00:09:18.122 "abort": true, 00:09:18.122 "seek_hole": false, 00:09:18.122 "seek_data": false, 00:09:18.122 "copy": true, 00:09:18.122 "nvme_iov_md": false 00:09:18.122 }, 00:09:18.122 "memory_domains": [ 00:09:18.122 { 00:09:18.122 "dma_device_id": "system", 00:09:18.122 "dma_device_type": 1 00:09:18.122 }, 00:09:18.122 { 00:09:18.122 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:18.122 "dma_device_type": 2 00:09:18.122 } 00:09:18.122 ], 00:09:18.122 "driver_specific": {} 00:09:18.122 } 00:09:18.122 ] 00:09:18.122 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.122 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:18.122 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:18.122 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:18.122 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:18.122 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:18.122 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:18.122 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:18.122 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:18.122 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:18.122 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:18.122 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:18.122 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:18.122 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:18.122 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.122 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.122 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.122 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:18.122 "name": "Existed_Raid", 00:09:18.122 "uuid": "8336f21e-ed91-424f-89f7-f0fe09684fbf", 00:09:18.122 "strip_size_kb": 0, 00:09:18.122 "state": "configuring", 00:09:18.122 "raid_level": "raid1", 00:09:18.122 "superblock": true, 00:09:18.122 "num_base_bdevs": 3, 00:09:18.122 "num_base_bdevs_discovered": 1, 00:09:18.122 "num_base_bdevs_operational": 3, 00:09:18.122 "base_bdevs_list": [ 00:09:18.122 { 00:09:18.122 "name": "BaseBdev1", 00:09:18.122 "uuid": "f17f1bfa-8f3d-46a6-8c64-428af4dc5e45", 00:09:18.122 "is_configured": true, 00:09:18.122 "data_offset": 2048, 00:09:18.122 "data_size": 63488 00:09:18.122 }, 00:09:18.122 { 00:09:18.122 "name": "BaseBdev2", 00:09:18.122 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:18.122 "is_configured": false, 00:09:18.122 "data_offset": 0, 00:09:18.122 "data_size": 0 00:09:18.122 }, 00:09:18.122 { 00:09:18.122 "name": "BaseBdev3", 00:09:18.122 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:18.122 "is_configured": false, 00:09:18.122 "data_offset": 0, 00:09:18.122 "data_size": 0 00:09:18.122 } 00:09:18.122 ] 00:09:18.122 }' 00:09:18.122 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:18.122 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.692 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:18.692 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.692 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.692 [2024-11-27 15:13:46.509573] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:18.692 [2024-11-27 15:13:46.509653] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:09:18.692 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.692 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:18.692 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.692 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.692 [2024-11-27 15:13:46.521581] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:18.692 [2024-11-27 15:13:46.523503] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:18.692 [2024-11-27 15:13:46.523539] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:18.692 [2024-11-27 15:13:46.523565] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:18.692 [2024-11-27 15:13:46.523575] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:18.692 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.692 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:18.692 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:18.692 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:18.692 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:18.692 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:18.692 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:18.692 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:18.692 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:18.692 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:18.692 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:18.692 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:18.692 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:18.692 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:18.692 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:18.692 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.692 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.692 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.692 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:18.692 "name": "Existed_Raid", 00:09:18.692 "uuid": "d206ff87-de26-4ff5-bc00-2c12da554862", 00:09:18.692 "strip_size_kb": 0, 00:09:18.692 "state": "configuring", 00:09:18.692 "raid_level": "raid1", 00:09:18.692 "superblock": true, 00:09:18.692 "num_base_bdevs": 3, 00:09:18.692 "num_base_bdevs_discovered": 1, 00:09:18.692 "num_base_bdevs_operational": 3, 00:09:18.692 "base_bdevs_list": [ 00:09:18.692 { 00:09:18.692 "name": "BaseBdev1", 00:09:18.692 "uuid": "f17f1bfa-8f3d-46a6-8c64-428af4dc5e45", 00:09:18.692 "is_configured": true, 00:09:18.692 "data_offset": 2048, 00:09:18.692 "data_size": 63488 00:09:18.692 }, 00:09:18.692 { 00:09:18.692 "name": "BaseBdev2", 00:09:18.692 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:18.692 "is_configured": false, 00:09:18.692 "data_offset": 0, 00:09:18.692 "data_size": 0 00:09:18.692 }, 00:09:18.692 { 00:09:18.692 "name": "BaseBdev3", 00:09:18.692 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:18.692 "is_configured": false, 00:09:18.692 "data_offset": 0, 00:09:18.692 "data_size": 0 00:09:18.692 } 00:09:18.692 ] 00:09:18.692 }' 00:09:18.692 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:18.692 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.952 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:18.952 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.952 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.952 [2024-11-27 15:13:46.975891] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:18.952 BaseBdev2 00:09:18.952 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.952 15:13:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:18.952 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:18.952 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:18.952 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:18.952 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:18.952 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:18.952 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:18.952 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.952 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.952 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.952 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:18.952 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.952 15:13:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.952 [ 00:09:18.952 { 00:09:18.952 "name": "BaseBdev2", 00:09:18.952 "aliases": [ 00:09:18.952 "53188e59-cdd4-43db-8867-e981f2e23f43" 00:09:18.952 ], 00:09:18.952 "product_name": "Malloc disk", 00:09:18.952 "block_size": 512, 00:09:18.952 "num_blocks": 65536, 00:09:18.952 "uuid": "53188e59-cdd4-43db-8867-e981f2e23f43", 00:09:18.952 "assigned_rate_limits": { 00:09:18.952 "rw_ios_per_sec": 0, 00:09:18.952 "rw_mbytes_per_sec": 0, 00:09:18.952 "r_mbytes_per_sec": 0, 00:09:18.952 "w_mbytes_per_sec": 0 00:09:18.952 }, 00:09:18.952 "claimed": true, 00:09:18.952 "claim_type": "exclusive_write", 00:09:18.952 "zoned": false, 00:09:18.952 "supported_io_types": { 00:09:18.952 "read": true, 00:09:18.952 "write": true, 00:09:18.952 "unmap": true, 00:09:18.952 "flush": true, 00:09:18.952 "reset": true, 00:09:18.952 "nvme_admin": false, 00:09:18.952 "nvme_io": false, 00:09:18.952 "nvme_io_md": false, 00:09:18.952 "write_zeroes": true, 00:09:18.952 "zcopy": true, 00:09:18.952 "get_zone_info": false, 00:09:18.952 "zone_management": false, 00:09:18.952 "zone_append": false, 00:09:18.952 "compare": false, 00:09:18.952 "compare_and_write": false, 00:09:18.952 "abort": true, 00:09:18.952 "seek_hole": false, 00:09:18.952 "seek_data": false, 00:09:18.952 "copy": true, 00:09:18.952 "nvme_iov_md": false 00:09:18.952 }, 00:09:18.952 "memory_domains": [ 00:09:18.952 { 00:09:18.952 "dma_device_id": "system", 00:09:18.952 "dma_device_type": 1 00:09:18.952 }, 00:09:18.952 { 00:09:18.952 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:18.952 "dma_device_type": 2 00:09:18.952 } 00:09:18.952 ], 00:09:18.952 "driver_specific": {} 00:09:18.952 } 00:09:18.952 ] 00:09:18.952 15:13:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.952 15:13:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:18.952 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:18.952 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:18.952 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:18.952 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:18.952 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:18.952 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:18.952 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:18.952 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:18.952 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:18.952 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:18.952 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:18.952 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:18.952 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:18.952 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:18.952 15:13:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.952 15:13:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.952 15:13:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.952 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:18.952 "name": "Existed_Raid", 00:09:18.952 "uuid": "d206ff87-de26-4ff5-bc00-2c12da554862", 00:09:18.952 "strip_size_kb": 0, 00:09:18.952 "state": "configuring", 00:09:18.952 "raid_level": "raid1", 00:09:18.952 "superblock": true, 00:09:18.952 "num_base_bdevs": 3, 00:09:18.952 "num_base_bdevs_discovered": 2, 00:09:18.952 "num_base_bdevs_operational": 3, 00:09:18.952 "base_bdevs_list": [ 00:09:18.952 { 00:09:18.952 "name": "BaseBdev1", 00:09:18.952 "uuid": "f17f1bfa-8f3d-46a6-8c64-428af4dc5e45", 00:09:18.952 "is_configured": true, 00:09:18.952 "data_offset": 2048, 00:09:18.952 "data_size": 63488 00:09:18.952 }, 00:09:18.952 { 00:09:18.952 "name": "BaseBdev2", 00:09:18.952 "uuid": "53188e59-cdd4-43db-8867-e981f2e23f43", 00:09:18.952 "is_configured": true, 00:09:18.952 "data_offset": 2048, 00:09:18.952 "data_size": 63488 00:09:18.952 }, 00:09:18.952 { 00:09:18.952 "name": "BaseBdev3", 00:09:18.952 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:18.952 "is_configured": false, 00:09:18.952 "data_offset": 0, 00:09:18.952 "data_size": 0 00:09:18.952 } 00:09:18.952 ] 00:09:18.952 }' 00:09:18.952 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:18.952 15:13:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.530 [2024-11-27 15:13:47.447502] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:19.530 [2024-11-27 15:13:47.447759] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:09:19.530 [2024-11-27 15:13:47.447777] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:19.530 [2024-11-27 15:13:47.448073] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:09:19.530 [2024-11-27 15:13:47.448221] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:09:19.530 [2024-11-27 15:13:47.448239] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:09:19.530 BaseBdev3 00:09:19.530 [2024-11-27 15:13:47.448380] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.530 [ 00:09:19.530 { 00:09:19.530 "name": "BaseBdev3", 00:09:19.530 "aliases": [ 00:09:19.530 "a887b762-53a7-476f-87cd-fd983c005e65" 00:09:19.530 ], 00:09:19.530 "product_name": "Malloc disk", 00:09:19.530 "block_size": 512, 00:09:19.530 "num_blocks": 65536, 00:09:19.530 "uuid": "a887b762-53a7-476f-87cd-fd983c005e65", 00:09:19.530 "assigned_rate_limits": { 00:09:19.530 "rw_ios_per_sec": 0, 00:09:19.530 "rw_mbytes_per_sec": 0, 00:09:19.530 "r_mbytes_per_sec": 0, 00:09:19.530 "w_mbytes_per_sec": 0 00:09:19.530 }, 00:09:19.530 "claimed": true, 00:09:19.530 "claim_type": "exclusive_write", 00:09:19.530 "zoned": false, 00:09:19.530 "supported_io_types": { 00:09:19.530 "read": true, 00:09:19.530 "write": true, 00:09:19.530 "unmap": true, 00:09:19.530 "flush": true, 00:09:19.530 "reset": true, 00:09:19.530 "nvme_admin": false, 00:09:19.530 "nvme_io": false, 00:09:19.530 "nvme_io_md": false, 00:09:19.530 "write_zeroes": true, 00:09:19.530 "zcopy": true, 00:09:19.530 "get_zone_info": false, 00:09:19.530 "zone_management": false, 00:09:19.530 "zone_append": false, 00:09:19.530 "compare": false, 00:09:19.530 "compare_and_write": false, 00:09:19.530 "abort": true, 00:09:19.530 "seek_hole": false, 00:09:19.530 "seek_data": false, 00:09:19.530 "copy": true, 00:09:19.530 "nvme_iov_md": false 00:09:19.530 }, 00:09:19.530 "memory_domains": [ 00:09:19.530 { 00:09:19.530 "dma_device_id": "system", 00:09:19.530 "dma_device_type": 1 00:09:19.530 }, 00:09:19.530 { 00:09:19.530 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:19.530 "dma_device_type": 2 00:09:19.530 } 00:09:19.530 ], 00:09:19.530 "driver_specific": {} 00:09:19.530 } 00:09:19.530 ] 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:19.530 "name": "Existed_Raid", 00:09:19.530 "uuid": "d206ff87-de26-4ff5-bc00-2c12da554862", 00:09:19.530 "strip_size_kb": 0, 00:09:19.530 "state": "online", 00:09:19.530 "raid_level": "raid1", 00:09:19.530 "superblock": true, 00:09:19.530 "num_base_bdevs": 3, 00:09:19.530 "num_base_bdevs_discovered": 3, 00:09:19.530 "num_base_bdevs_operational": 3, 00:09:19.530 "base_bdevs_list": [ 00:09:19.530 { 00:09:19.530 "name": "BaseBdev1", 00:09:19.530 "uuid": "f17f1bfa-8f3d-46a6-8c64-428af4dc5e45", 00:09:19.530 "is_configured": true, 00:09:19.530 "data_offset": 2048, 00:09:19.530 "data_size": 63488 00:09:19.530 }, 00:09:19.530 { 00:09:19.530 "name": "BaseBdev2", 00:09:19.530 "uuid": "53188e59-cdd4-43db-8867-e981f2e23f43", 00:09:19.530 "is_configured": true, 00:09:19.530 "data_offset": 2048, 00:09:19.530 "data_size": 63488 00:09:19.530 }, 00:09:19.530 { 00:09:19.530 "name": "BaseBdev3", 00:09:19.530 "uuid": "a887b762-53a7-476f-87cd-fd983c005e65", 00:09:19.530 "is_configured": true, 00:09:19.530 "data_offset": 2048, 00:09:19.530 "data_size": 63488 00:09:19.530 } 00:09:19.530 ] 00:09:19.530 }' 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:19.530 15:13:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.807 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:19.807 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:19.807 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:19.807 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:19.807 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:19.807 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:19.807 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:19.807 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:19.807 15:13:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.807 15:13:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.807 [2024-11-27 15:13:47.875183] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:19.807 15:13:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.807 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:19.807 "name": "Existed_Raid", 00:09:19.807 "aliases": [ 00:09:19.807 "d206ff87-de26-4ff5-bc00-2c12da554862" 00:09:19.807 ], 00:09:19.807 "product_name": "Raid Volume", 00:09:19.807 "block_size": 512, 00:09:19.807 "num_blocks": 63488, 00:09:19.807 "uuid": "d206ff87-de26-4ff5-bc00-2c12da554862", 00:09:19.807 "assigned_rate_limits": { 00:09:19.807 "rw_ios_per_sec": 0, 00:09:19.807 "rw_mbytes_per_sec": 0, 00:09:19.807 "r_mbytes_per_sec": 0, 00:09:19.807 "w_mbytes_per_sec": 0 00:09:19.807 }, 00:09:19.807 "claimed": false, 00:09:19.807 "zoned": false, 00:09:19.807 "supported_io_types": { 00:09:19.807 "read": true, 00:09:19.807 "write": true, 00:09:19.807 "unmap": false, 00:09:19.807 "flush": false, 00:09:19.807 "reset": true, 00:09:19.807 "nvme_admin": false, 00:09:19.807 "nvme_io": false, 00:09:19.807 "nvme_io_md": false, 00:09:19.807 "write_zeroes": true, 00:09:19.807 "zcopy": false, 00:09:19.807 "get_zone_info": false, 00:09:19.807 "zone_management": false, 00:09:19.807 "zone_append": false, 00:09:19.807 "compare": false, 00:09:19.807 "compare_and_write": false, 00:09:19.807 "abort": false, 00:09:19.807 "seek_hole": false, 00:09:19.807 "seek_data": false, 00:09:19.807 "copy": false, 00:09:19.807 "nvme_iov_md": false 00:09:19.807 }, 00:09:19.807 "memory_domains": [ 00:09:19.807 { 00:09:19.807 "dma_device_id": "system", 00:09:19.807 "dma_device_type": 1 00:09:19.807 }, 00:09:19.807 { 00:09:19.807 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:19.807 "dma_device_type": 2 00:09:19.807 }, 00:09:19.807 { 00:09:19.807 "dma_device_id": "system", 00:09:19.807 "dma_device_type": 1 00:09:19.807 }, 00:09:19.807 { 00:09:19.807 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:19.807 "dma_device_type": 2 00:09:19.807 }, 00:09:19.807 { 00:09:19.807 "dma_device_id": "system", 00:09:19.807 "dma_device_type": 1 00:09:19.807 }, 00:09:19.807 { 00:09:19.807 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:19.807 "dma_device_type": 2 00:09:19.807 } 00:09:19.807 ], 00:09:19.807 "driver_specific": { 00:09:19.807 "raid": { 00:09:19.807 "uuid": "d206ff87-de26-4ff5-bc00-2c12da554862", 00:09:19.807 "strip_size_kb": 0, 00:09:19.807 "state": "online", 00:09:19.807 "raid_level": "raid1", 00:09:19.807 "superblock": true, 00:09:19.807 "num_base_bdevs": 3, 00:09:19.807 "num_base_bdevs_discovered": 3, 00:09:19.807 "num_base_bdevs_operational": 3, 00:09:19.807 "base_bdevs_list": [ 00:09:19.807 { 00:09:19.807 "name": "BaseBdev1", 00:09:19.807 "uuid": "f17f1bfa-8f3d-46a6-8c64-428af4dc5e45", 00:09:19.807 "is_configured": true, 00:09:19.807 "data_offset": 2048, 00:09:19.807 "data_size": 63488 00:09:19.807 }, 00:09:19.807 { 00:09:19.807 "name": "BaseBdev2", 00:09:19.807 "uuid": "53188e59-cdd4-43db-8867-e981f2e23f43", 00:09:19.807 "is_configured": true, 00:09:19.807 "data_offset": 2048, 00:09:19.807 "data_size": 63488 00:09:19.807 }, 00:09:19.807 { 00:09:19.807 "name": "BaseBdev3", 00:09:19.807 "uuid": "a887b762-53a7-476f-87cd-fd983c005e65", 00:09:19.807 "is_configured": true, 00:09:19.807 "data_offset": 2048, 00:09:19.807 "data_size": 63488 00:09:19.807 } 00:09:19.807 ] 00:09:19.807 } 00:09:19.807 } 00:09:19.807 }' 00:09:20.068 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:20.068 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:20.068 BaseBdev2 00:09:20.068 BaseBdev3' 00:09:20.068 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:20.068 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:20.068 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:20.068 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:20.068 15:13:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:20.068 15:13:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.068 15:13:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.068 [2024-11-27 15:13:48.126511] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.068 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.328 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:20.328 "name": "Existed_Raid", 00:09:20.328 "uuid": "d206ff87-de26-4ff5-bc00-2c12da554862", 00:09:20.328 "strip_size_kb": 0, 00:09:20.328 "state": "online", 00:09:20.328 "raid_level": "raid1", 00:09:20.328 "superblock": true, 00:09:20.328 "num_base_bdevs": 3, 00:09:20.328 "num_base_bdevs_discovered": 2, 00:09:20.328 "num_base_bdevs_operational": 2, 00:09:20.328 "base_bdevs_list": [ 00:09:20.328 { 00:09:20.328 "name": null, 00:09:20.328 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:20.328 "is_configured": false, 00:09:20.328 "data_offset": 0, 00:09:20.328 "data_size": 63488 00:09:20.328 }, 00:09:20.328 { 00:09:20.328 "name": "BaseBdev2", 00:09:20.328 "uuid": "53188e59-cdd4-43db-8867-e981f2e23f43", 00:09:20.328 "is_configured": true, 00:09:20.328 "data_offset": 2048, 00:09:20.328 "data_size": 63488 00:09:20.328 }, 00:09:20.328 { 00:09:20.328 "name": "BaseBdev3", 00:09:20.328 "uuid": "a887b762-53a7-476f-87cd-fd983c005e65", 00:09:20.328 "is_configured": true, 00:09:20.328 "data_offset": 2048, 00:09:20.328 "data_size": 63488 00:09:20.328 } 00:09:20.328 ] 00:09:20.328 }' 00:09:20.328 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:20.328 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.588 [2024-11-27 15:13:48.593395] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.588 [2024-11-27 15:13:48.660397] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:20.588 [2024-11-27 15:13:48.660548] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:20.588 [2024-11-27 15:13:48.671990] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:20.588 [2024-11-27 15:13:48.672097] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:20.588 [2024-11-27 15:13:48.672142] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.588 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.849 BaseBdev2 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.849 [ 00:09:20.849 { 00:09:20.849 "name": "BaseBdev2", 00:09:20.849 "aliases": [ 00:09:20.849 "4eb7842e-4f5d-4254-a806-8eb513340583" 00:09:20.849 ], 00:09:20.849 "product_name": "Malloc disk", 00:09:20.849 "block_size": 512, 00:09:20.849 "num_blocks": 65536, 00:09:20.849 "uuid": "4eb7842e-4f5d-4254-a806-8eb513340583", 00:09:20.849 "assigned_rate_limits": { 00:09:20.849 "rw_ios_per_sec": 0, 00:09:20.849 "rw_mbytes_per_sec": 0, 00:09:20.849 "r_mbytes_per_sec": 0, 00:09:20.849 "w_mbytes_per_sec": 0 00:09:20.849 }, 00:09:20.849 "claimed": false, 00:09:20.849 "zoned": false, 00:09:20.849 "supported_io_types": { 00:09:20.849 "read": true, 00:09:20.849 "write": true, 00:09:20.849 "unmap": true, 00:09:20.849 "flush": true, 00:09:20.849 "reset": true, 00:09:20.849 "nvme_admin": false, 00:09:20.849 "nvme_io": false, 00:09:20.849 "nvme_io_md": false, 00:09:20.849 "write_zeroes": true, 00:09:20.849 "zcopy": true, 00:09:20.849 "get_zone_info": false, 00:09:20.849 "zone_management": false, 00:09:20.849 "zone_append": false, 00:09:20.849 "compare": false, 00:09:20.849 "compare_and_write": false, 00:09:20.849 "abort": true, 00:09:20.849 "seek_hole": false, 00:09:20.849 "seek_data": false, 00:09:20.849 "copy": true, 00:09:20.849 "nvme_iov_md": false 00:09:20.849 }, 00:09:20.849 "memory_domains": [ 00:09:20.849 { 00:09:20.849 "dma_device_id": "system", 00:09:20.849 "dma_device_type": 1 00:09:20.849 }, 00:09:20.849 { 00:09:20.849 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:20.849 "dma_device_type": 2 00:09:20.849 } 00:09:20.849 ], 00:09:20.849 "driver_specific": {} 00:09:20.849 } 00:09:20.849 ] 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.849 BaseBdev3 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.849 [ 00:09:20.849 { 00:09:20.849 "name": "BaseBdev3", 00:09:20.849 "aliases": [ 00:09:20.849 "2aded048-ac16-417d-aa08-7719100fd889" 00:09:20.849 ], 00:09:20.849 "product_name": "Malloc disk", 00:09:20.849 "block_size": 512, 00:09:20.849 "num_blocks": 65536, 00:09:20.849 "uuid": "2aded048-ac16-417d-aa08-7719100fd889", 00:09:20.849 "assigned_rate_limits": { 00:09:20.849 "rw_ios_per_sec": 0, 00:09:20.849 "rw_mbytes_per_sec": 0, 00:09:20.849 "r_mbytes_per_sec": 0, 00:09:20.849 "w_mbytes_per_sec": 0 00:09:20.849 }, 00:09:20.849 "claimed": false, 00:09:20.849 "zoned": false, 00:09:20.849 "supported_io_types": { 00:09:20.849 "read": true, 00:09:20.849 "write": true, 00:09:20.849 "unmap": true, 00:09:20.849 "flush": true, 00:09:20.849 "reset": true, 00:09:20.849 "nvme_admin": false, 00:09:20.849 "nvme_io": false, 00:09:20.849 "nvme_io_md": false, 00:09:20.849 "write_zeroes": true, 00:09:20.849 "zcopy": true, 00:09:20.849 "get_zone_info": false, 00:09:20.849 "zone_management": false, 00:09:20.849 "zone_append": false, 00:09:20.849 "compare": false, 00:09:20.849 "compare_and_write": false, 00:09:20.849 "abort": true, 00:09:20.849 "seek_hole": false, 00:09:20.849 "seek_data": false, 00:09:20.849 "copy": true, 00:09:20.849 "nvme_iov_md": false 00:09:20.849 }, 00:09:20.849 "memory_domains": [ 00:09:20.849 { 00:09:20.849 "dma_device_id": "system", 00:09:20.849 "dma_device_type": 1 00:09:20.849 }, 00:09:20.849 { 00:09:20.849 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:20.849 "dma_device_type": 2 00:09:20.849 } 00:09:20.849 ], 00:09:20.849 "driver_specific": {} 00:09:20.849 } 00:09:20.849 ] 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:20.849 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:20.850 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.850 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.850 [2024-11-27 15:13:48.823109] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:20.850 [2024-11-27 15:13:48.823232] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:20.850 [2024-11-27 15:13:48.823269] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:20.850 [2024-11-27 15:13:48.825095] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:20.850 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.850 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:20.850 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:20.850 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:20.850 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:20.850 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:20.850 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:20.850 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:20.850 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:20.850 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:20.850 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:20.850 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:20.850 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:20.850 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.850 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.850 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.850 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:20.850 "name": "Existed_Raid", 00:09:20.850 "uuid": "33e777d9-b7ff-4c4b-9e9a-e02102bee68a", 00:09:20.850 "strip_size_kb": 0, 00:09:20.850 "state": "configuring", 00:09:20.850 "raid_level": "raid1", 00:09:20.850 "superblock": true, 00:09:20.850 "num_base_bdevs": 3, 00:09:20.850 "num_base_bdevs_discovered": 2, 00:09:20.850 "num_base_bdevs_operational": 3, 00:09:20.850 "base_bdevs_list": [ 00:09:20.850 { 00:09:20.850 "name": "BaseBdev1", 00:09:20.850 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:20.850 "is_configured": false, 00:09:20.850 "data_offset": 0, 00:09:20.850 "data_size": 0 00:09:20.850 }, 00:09:20.850 { 00:09:20.850 "name": "BaseBdev2", 00:09:20.850 "uuid": "4eb7842e-4f5d-4254-a806-8eb513340583", 00:09:20.850 "is_configured": true, 00:09:20.850 "data_offset": 2048, 00:09:20.850 "data_size": 63488 00:09:20.850 }, 00:09:20.850 { 00:09:20.850 "name": "BaseBdev3", 00:09:20.850 "uuid": "2aded048-ac16-417d-aa08-7719100fd889", 00:09:20.850 "is_configured": true, 00:09:20.850 "data_offset": 2048, 00:09:20.850 "data_size": 63488 00:09:20.850 } 00:09:20.850 ] 00:09:20.850 }' 00:09:20.850 15:13:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:20.850 15:13:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:21.418 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:21.418 15:13:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.418 15:13:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:21.418 [2024-11-27 15:13:49.278364] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:21.418 15:13:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.418 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:21.418 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:21.418 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:21.418 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:21.418 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:21.418 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:21.418 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:21.418 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:21.418 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:21.418 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:21.418 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:21.418 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:21.418 15:13:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.418 15:13:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:21.418 15:13:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.418 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:21.418 "name": "Existed_Raid", 00:09:21.418 "uuid": "33e777d9-b7ff-4c4b-9e9a-e02102bee68a", 00:09:21.418 "strip_size_kb": 0, 00:09:21.418 "state": "configuring", 00:09:21.418 "raid_level": "raid1", 00:09:21.418 "superblock": true, 00:09:21.418 "num_base_bdevs": 3, 00:09:21.418 "num_base_bdevs_discovered": 1, 00:09:21.418 "num_base_bdevs_operational": 3, 00:09:21.418 "base_bdevs_list": [ 00:09:21.418 { 00:09:21.418 "name": "BaseBdev1", 00:09:21.418 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:21.418 "is_configured": false, 00:09:21.418 "data_offset": 0, 00:09:21.418 "data_size": 0 00:09:21.418 }, 00:09:21.418 { 00:09:21.418 "name": null, 00:09:21.418 "uuid": "4eb7842e-4f5d-4254-a806-8eb513340583", 00:09:21.418 "is_configured": false, 00:09:21.418 "data_offset": 0, 00:09:21.418 "data_size": 63488 00:09:21.418 }, 00:09:21.418 { 00:09:21.418 "name": "BaseBdev3", 00:09:21.418 "uuid": "2aded048-ac16-417d-aa08-7719100fd889", 00:09:21.418 "is_configured": true, 00:09:21.418 "data_offset": 2048, 00:09:21.418 "data_size": 63488 00:09:21.418 } 00:09:21.418 ] 00:09:21.418 }' 00:09:21.418 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:21.418 15:13:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:21.677 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:21.677 15:13:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.677 15:13:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:21.677 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:21.677 15:13:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.677 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:21.677 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:21.677 15:13:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.677 15:13:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:21.677 [2024-11-27 15:13:49.768501] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:21.677 BaseBdev1 00:09:21.677 15:13:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.677 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:21.677 15:13:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:21.677 15:13:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:21.677 15:13:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:21.677 15:13:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:21.677 15:13:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:21.677 15:13:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:21.677 15:13:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.677 15:13:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:21.677 15:13:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.677 15:13:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:21.935 15:13:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.935 15:13:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:21.935 [ 00:09:21.935 { 00:09:21.935 "name": "BaseBdev1", 00:09:21.935 "aliases": [ 00:09:21.935 "84da1d46-e0a9-4138-9346-49319a867567" 00:09:21.935 ], 00:09:21.935 "product_name": "Malloc disk", 00:09:21.935 "block_size": 512, 00:09:21.935 "num_blocks": 65536, 00:09:21.935 "uuid": "84da1d46-e0a9-4138-9346-49319a867567", 00:09:21.935 "assigned_rate_limits": { 00:09:21.935 "rw_ios_per_sec": 0, 00:09:21.935 "rw_mbytes_per_sec": 0, 00:09:21.935 "r_mbytes_per_sec": 0, 00:09:21.935 "w_mbytes_per_sec": 0 00:09:21.935 }, 00:09:21.935 "claimed": true, 00:09:21.935 "claim_type": "exclusive_write", 00:09:21.935 "zoned": false, 00:09:21.935 "supported_io_types": { 00:09:21.935 "read": true, 00:09:21.935 "write": true, 00:09:21.935 "unmap": true, 00:09:21.935 "flush": true, 00:09:21.935 "reset": true, 00:09:21.935 "nvme_admin": false, 00:09:21.935 "nvme_io": false, 00:09:21.935 "nvme_io_md": false, 00:09:21.935 "write_zeroes": true, 00:09:21.935 "zcopy": true, 00:09:21.935 "get_zone_info": false, 00:09:21.935 "zone_management": false, 00:09:21.935 "zone_append": false, 00:09:21.935 "compare": false, 00:09:21.935 "compare_and_write": false, 00:09:21.935 "abort": true, 00:09:21.935 "seek_hole": false, 00:09:21.935 "seek_data": false, 00:09:21.935 "copy": true, 00:09:21.935 "nvme_iov_md": false 00:09:21.935 }, 00:09:21.935 "memory_domains": [ 00:09:21.935 { 00:09:21.935 "dma_device_id": "system", 00:09:21.935 "dma_device_type": 1 00:09:21.935 }, 00:09:21.935 { 00:09:21.935 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:21.935 "dma_device_type": 2 00:09:21.935 } 00:09:21.935 ], 00:09:21.935 "driver_specific": {} 00:09:21.935 } 00:09:21.935 ] 00:09:21.935 15:13:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.935 15:13:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:21.935 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:21.935 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:21.935 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:21.935 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:21.935 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:21.935 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:21.935 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:21.935 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:21.935 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:21.936 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:21.936 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:21.936 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:21.936 15:13:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.936 15:13:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:21.936 15:13:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.936 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:21.936 "name": "Existed_Raid", 00:09:21.936 "uuid": "33e777d9-b7ff-4c4b-9e9a-e02102bee68a", 00:09:21.936 "strip_size_kb": 0, 00:09:21.936 "state": "configuring", 00:09:21.936 "raid_level": "raid1", 00:09:21.936 "superblock": true, 00:09:21.936 "num_base_bdevs": 3, 00:09:21.936 "num_base_bdevs_discovered": 2, 00:09:21.936 "num_base_bdevs_operational": 3, 00:09:21.936 "base_bdevs_list": [ 00:09:21.936 { 00:09:21.936 "name": "BaseBdev1", 00:09:21.936 "uuid": "84da1d46-e0a9-4138-9346-49319a867567", 00:09:21.936 "is_configured": true, 00:09:21.936 "data_offset": 2048, 00:09:21.936 "data_size": 63488 00:09:21.936 }, 00:09:21.936 { 00:09:21.936 "name": null, 00:09:21.936 "uuid": "4eb7842e-4f5d-4254-a806-8eb513340583", 00:09:21.936 "is_configured": false, 00:09:21.936 "data_offset": 0, 00:09:21.936 "data_size": 63488 00:09:21.936 }, 00:09:21.936 { 00:09:21.936 "name": "BaseBdev3", 00:09:21.936 "uuid": "2aded048-ac16-417d-aa08-7719100fd889", 00:09:21.936 "is_configured": true, 00:09:21.936 "data_offset": 2048, 00:09:21.936 "data_size": 63488 00:09:21.936 } 00:09:21.936 ] 00:09:21.936 }' 00:09:21.936 15:13:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:21.936 15:13:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:22.195 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:22.195 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:22.195 15:13:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.195 15:13:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:22.195 15:13:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.455 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:22.455 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:22.455 15:13:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.455 15:13:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:22.455 [2024-11-27 15:13:50.319601] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:22.455 15:13:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.455 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:22.456 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:22.456 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:22.456 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:22.456 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:22.456 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:22.456 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:22.456 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:22.456 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:22.456 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:22.456 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:22.456 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:22.456 15:13:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.456 15:13:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:22.456 15:13:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.456 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:22.456 "name": "Existed_Raid", 00:09:22.456 "uuid": "33e777d9-b7ff-4c4b-9e9a-e02102bee68a", 00:09:22.456 "strip_size_kb": 0, 00:09:22.456 "state": "configuring", 00:09:22.456 "raid_level": "raid1", 00:09:22.456 "superblock": true, 00:09:22.456 "num_base_bdevs": 3, 00:09:22.456 "num_base_bdevs_discovered": 1, 00:09:22.456 "num_base_bdevs_operational": 3, 00:09:22.456 "base_bdevs_list": [ 00:09:22.456 { 00:09:22.456 "name": "BaseBdev1", 00:09:22.456 "uuid": "84da1d46-e0a9-4138-9346-49319a867567", 00:09:22.456 "is_configured": true, 00:09:22.456 "data_offset": 2048, 00:09:22.456 "data_size": 63488 00:09:22.456 }, 00:09:22.456 { 00:09:22.456 "name": null, 00:09:22.456 "uuid": "4eb7842e-4f5d-4254-a806-8eb513340583", 00:09:22.456 "is_configured": false, 00:09:22.456 "data_offset": 0, 00:09:22.456 "data_size": 63488 00:09:22.456 }, 00:09:22.456 { 00:09:22.456 "name": null, 00:09:22.456 "uuid": "2aded048-ac16-417d-aa08-7719100fd889", 00:09:22.456 "is_configured": false, 00:09:22.456 "data_offset": 0, 00:09:22.456 "data_size": 63488 00:09:22.456 } 00:09:22.456 ] 00:09:22.456 }' 00:09:22.456 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:22.456 15:13:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:22.715 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:22.715 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:22.715 15:13:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.715 15:13:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:22.715 15:13:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.715 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:22.715 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:22.715 15:13:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.715 15:13:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:22.715 [2024-11-27 15:13:50.774818] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:22.715 15:13:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.715 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:22.715 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:22.715 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:22.715 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:22.715 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:22.715 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:22.715 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:22.715 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:22.715 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:22.715 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:22.715 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:22.715 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:22.715 15:13:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.715 15:13:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:22.715 15:13:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.975 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:22.975 "name": "Existed_Raid", 00:09:22.975 "uuid": "33e777d9-b7ff-4c4b-9e9a-e02102bee68a", 00:09:22.975 "strip_size_kb": 0, 00:09:22.975 "state": "configuring", 00:09:22.975 "raid_level": "raid1", 00:09:22.975 "superblock": true, 00:09:22.975 "num_base_bdevs": 3, 00:09:22.975 "num_base_bdevs_discovered": 2, 00:09:22.975 "num_base_bdevs_operational": 3, 00:09:22.975 "base_bdevs_list": [ 00:09:22.975 { 00:09:22.975 "name": "BaseBdev1", 00:09:22.975 "uuid": "84da1d46-e0a9-4138-9346-49319a867567", 00:09:22.975 "is_configured": true, 00:09:22.975 "data_offset": 2048, 00:09:22.975 "data_size": 63488 00:09:22.975 }, 00:09:22.975 { 00:09:22.975 "name": null, 00:09:22.975 "uuid": "4eb7842e-4f5d-4254-a806-8eb513340583", 00:09:22.975 "is_configured": false, 00:09:22.975 "data_offset": 0, 00:09:22.975 "data_size": 63488 00:09:22.975 }, 00:09:22.975 { 00:09:22.975 "name": "BaseBdev3", 00:09:22.975 "uuid": "2aded048-ac16-417d-aa08-7719100fd889", 00:09:22.975 "is_configured": true, 00:09:22.975 "data_offset": 2048, 00:09:22.975 "data_size": 63488 00:09:22.975 } 00:09:22.975 ] 00:09:22.975 }' 00:09:22.975 15:13:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:22.975 15:13:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:23.235 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:23.235 15:13:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.235 15:13:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:23.235 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:23.235 15:13:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.235 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:23.235 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:23.235 15:13:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.235 15:13:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:23.235 [2024-11-27 15:13:51.290004] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:23.235 15:13:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.235 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:23.235 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:23.235 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:23.235 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:23.235 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:23.235 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:23.235 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:23.235 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:23.235 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:23.235 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:23.235 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:23.235 15:13:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.235 15:13:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:23.235 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:23.235 15:13:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.495 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:23.495 "name": "Existed_Raid", 00:09:23.495 "uuid": "33e777d9-b7ff-4c4b-9e9a-e02102bee68a", 00:09:23.495 "strip_size_kb": 0, 00:09:23.495 "state": "configuring", 00:09:23.495 "raid_level": "raid1", 00:09:23.495 "superblock": true, 00:09:23.495 "num_base_bdevs": 3, 00:09:23.495 "num_base_bdevs_discovered": 1, 00:09:23.495 "num_base_bdevs_operational": 3, 00:09:23.495 "base_bdevs_list": [ 00:09:23.495 { 00:09:23.495 "name": null, 00:09:23.495 "uuid": "84da1d46-e0a9-4138-9346-49319a867567", 00:09:23.495 "is_configured": false, 00:09:23.495 "data_offset": 0, 00:09:23.495 "data_size": 63488 00:09:23.495 }, 00:09:23.495 { 00:09:23.495 "name": null, 00:09:23.495 "uuid": "4eb7842e-4f5d-4254-a806-8eb513340583", 00:09:23.495 "is_configured": false, 00:09:23.495 "data_offset": 0, 00:09:23.495 "data_size": 63488 00:09:23.495 }, 00:09:23.495 { 00:09:23.495 "name": "BaseBdev3", 00:09:23.495 "uuid": "2aded048-ac16-417d-aa08-7719100fd889", 00:09:23.495 "is_configured": true, 00:09:23.495 "data_offset": 2048, 00:09:23.495 "data_size": 63488 00:09:23.495 } 00:09:23.495 ] 00:09:23.495 }' 00:09:23.495 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:23.495 15:13:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:23.754 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:23.754 15:13:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.754 15:13:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:23.754 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:23.754 15:13:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.754 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:23.754 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:23.754 15:13:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.754 15:13:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:23.754 [2024-11-27 15:13:51.779673] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:23.754 15:13:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.754 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:23.754 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:23.754 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:23.754 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:23.755 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:23.755 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:23.755 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:23.755 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:23.755 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:23.755 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:23.755 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:23.755 15:13:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.755 15:13:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:23.755 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:23.755 15:13:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.755 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:23.755 "name": "Existed_Raid", 00:09:23.755 "uuid": "33e777d9-b7ff-4c4b-9e9a-e02102bee68a", 00:09:23.755 "strip_size_kb": 0, 00:09:23.755 "state": "configuring", 00:09:23.755 "raid_level": "raid1", 00:09:23.755 "superblock": true, 00:09:23.755 "num_base_bdevs": 3, 00:09:23.755 "num_base_bdevs_discovered": 2, 00:09:23.755 "num_base_bdevs_operational": 3, 00:09:23.755 "base_bdevs_list": [ 00:09:23.755 { 00:09:23.755 "name": null, 00:09:23.755 "uuid": "84da1d46-e0a9-4138-9346-49319a867567", 00:09:23.755 "is_configured": false, 00:09:23.755 "data_offset": 0, 00:09:23.755 "data_size": 63488 00:09:23.755 }, 00:09:23.755 { 00:09:23.755 "name": "BaseBdev2", 00:09:23.755 "uuid": "4eb7842e-4f5d-4254-a806-8eb513340583", 00:09:23.755 "is_configured": true, 00:09:23.755 "data_offset": 2048, 00:09:23.755 "data_size": 63488 00:09:23.755 }, 00:09:23.755 { 00:09:23.755 "name": "BaseBdev3", 00:09:23.755 "uuid": "2aded048-ac16-417d-aa08-7719100fd889", 00:09:23.755 "is_configured": true, 00:09:23.755 "data_offset": 2048, 00:09:23.755 "data_size": 63488 00:09:23.755 } 00:09:23.755 ] 00:09:23.755 }' 00:09:23.755 15:13:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:23.755 15:13:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 84da1d46-e0a9-4138-9346-49319a867567 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:24.324 [2024-11-27 15:13:52.301787] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:24.324 [2024-11-27 15:13:52.301970] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:09:24.324 [2024-11-27 15:13:52.301984] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:24.324 NewBaseBdev 00:09:24.324 [2024-11-27 15:13:52.302256] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:09:24.324 [2024-11-27 15:13:52.302379] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:09:24.324 [2024-11-27 15:13:52.302393] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006d00 00:09:24.324 [2024-11-27 15:13:52.302497] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:24.324 [ 00:09:24.324 { 00:09:24.324 "name": "NewBaseBdev", 00:09:24.324 "aliases": [ 00:09:24.324 "84da1d46-e0a9-4138-9346-49319a867567" 00:09:24.324 ], 00:09:24.324 "product_name": "Malloc disk", 00:09:24.324 "block_size": 512, 00:09:24.324 "num_blocks": 65536, 00:09:24.324 "uuid": "84da1d46-e0a9-4138-9346-49319a867567", 00:09:24.324 "assigned_rate_limits": { 00:09:24.324 "rw_ios_per_sec": 0, 00:09:24.324 "rw_mbytes_per_sec": 0, 00:09:24.324 "r_mbytes_per_sec": 0, 00:09:24.324 "w_mbytes_per_sec": 0 00:09:24.324 }, 00:09:24.324 "claimed": true, 00:09:24.324 "claim_type": "exclusive_write", 00:09:24.324 "zoned": false, 00:09:24.324 "supported_io_types": { 00:09:24.324 "read": true, 00:09:24.324 "write": true, 00:09:24.324 "unmap": true, 00:09:24.324 "flush": true, 00:09:24.324 "reset": true, 00:09:24.324 "nvme_admin": false, 00:09:24.324 "nvme_io": false, 00:09:24.324 "nvme_io_md": false, 00:09:24.324 "write_zeroes": true, 00:09:24.324 "zcopy": true, 00:09:24.324 "get_zone_info": false, 00:09:24.324 "zone_management": false, 00:09:24.324 "zone_append": false, 00:09:24.324 "compare": false, 00:09:24.324 "compare_and_write": false, 00:09:24.324 "abort": true, 00:09:24.324 "seek_hole": false, 00:09:24.324 "seek_data": false, 00:09:24.324 "copy": true, 00:09:24.324 "nvme_iov_md": false 00:09:24.324 }, 00:09:24.324 "memory_domains": [ 00:09:24.324 { 00:09:24.324 "dma_device_id": "system", 00:09:24.324 "dma_device_type": 1 00:09:24.324 }, 00:09:24.324 { 00:09:24.324 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:24.324 "dma_device_type": 2 00:09:24.324 } 00:09:24.324 ], 00:09:24.324 "driver_specific": {} 00:09:24.324 } 00:09:24.324 ] 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:24.324 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:24.324 "name": "Existed_Raid", 00:09:24.324 "uuid": "33e777d9-b7ff-4c4b-9e9a-e02102bee68a", 00:09:24.324 "strip_size_kb": 0, 00:09:24.324 "state": "online", 00:09:24.324 "raid_level": "raid1", 00:09:24.324 "superblock": true, 00:09:24.324 "num_base_bdevs": 3, 00:09:24.324 "num_base_bdevs_discovered": 3, 00:09:24.324 "num_base_bdevs_operational": 3, 00:09:24.324 "base_bdevs_list": [ 00:09:24.324 { 00:09:24.324 "name": "NewBaseBdev", 00:09:24.324 "uuid": "84da1d46-e0a9-4138-9346-49319a867567", 00:09:24.324 "is_configured": true, 00:09:24.324 "data_offset": 2048, 00:09:24.324 "data_size": 63488 00:09:24.324 }, 00:09:24.324 { 00:09:24.324 "name": "BaseBdev2", 00:09:24.324 "uuid": "4eb7842e-4f5d-4254-a806-8eb513340583", 00:09:24.324 "is_configured": true, 00:09:24.324 "data_offset": 2048, 00:09:24.324 "data_size": 63488 00:09:24.324 }, 00:09:24.324 { 00:09:24.324 "name": "BaseBdev3", 00:09:24.325 "uuid": "2aded048-ac16-417d-aa08-7719100fd889", 00:09:24.325 "is_configured": true, 00:09:24.325 "data_offset": 2048, 00:09:24.325 "data_size": 63488 00:09:24.325 } 00:09:24.325 ] 00:09:24.325 }' 00:09:24.325 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:24.325 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:24.894 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:24.894 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:24.894 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:24.894 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:24.894 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:24.894 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:24.894 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:24.894 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:24.894 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:24.894 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:24.894 [2024-11-27 15:13:52.793293] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:24.894 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:24.894 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:24.894 "name": "Existed_Raid", 00:09:24.894 "aliases": [ 00:09:24.894 "33e777d9-b7ff-4c4b-9e9a-e02102bee68a" 00:09:24.894 ], 00:09:24.894 "product_name": "Raid Volume", 00:09:24.894 "block_size": 512, 00:09:24.894 "num_blocks": 63488, 00:09:24.894 "uuid": "33e777d9-b7ff-4c4b-9e9a-e02102bee68a", 00:09:24.894 "assigned_rate_limits": { 00:09:24.894 "rw_ios_per_sec": 0, 00:09:24.894 "rw_mbytes_per_sec": 0, 00:09:24.894 "r_mbytes_per_sec": 0, 00:09:24.894 "w_mbytes_per_sec": 0 00:09:24.894 }, 00:09:24.894 "claimed": false, 00:09:24.894 "zoned": false, 00:09:24.894 "supported_io_types": { 00:09:24.894 "read": true, 00:09:24.894 "write": true, 00:09:24.894 "unmap": false, 00:09:24.894 "flush": false, 00:09:24.894 "reset": true, 00:09:24.894 "nvme_admin": false, 00:09:24.894 "nvme_io": false, 00:09:24.894 "nvme_io_md": false, 00:09:24.894 "write_zeroes": true, 00:09:24.894 "zcopy": false, 00:09:24.894 "get_zone_info": false, 00:09:24.894 "zone_management": false, 00:09:24.894 "zone_append": false, 00:09:24.894 "compare": false, 00:09:24.894 "compare_and_write": false, 00:09:24.894 "abort": false, 00:09:24.894 "seek_hole": false, 00:09:24.894 "seek_data": false, 00:09:24.894 "copy": false, 00:09:24.894 "nvme_iov_md": false 00:09:24.894 }, 00:09:24.894 "memory_domains": [ 00:09:24.894 { 00:09:24.894 "dma_device_id": "system", 00:09:24.894 "dma_device_type": 1 00:09:24.894 }, 00:09:24.894 { 00:09:24.894 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:24.894 "dma_device_type": 2 00:09:24.894 }, 00:09:24.894 { 00:09:24.894 "dma_device_id": "system", 00:09:24.894 "dma_device_type": 1 00:09:24.894 }, 00:09:24.894 { 00:09:24.894 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:24.894 "dma_device_type": 2 00:09:24.894 }, 00:09:24.894 { 00:09:24.894 "dma_device_id": "system", 00:09:24.894 "dma_device_type": 1 00:09:24.894 }, 00:09:24.894 { 00:09:24.894 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:24.894 "dma_device_type": 2 00:09:24.895 } 00:09:24.895 ], 00:09:24.895 "driver_specific": { 00:09:24.895 "raid": { 00:09:24.895 "uuid": "33e777d9-b7ff-4c4b-9e9a-e02102bee68a", 00:09:24.895 "strip_size_kb": 0, 00:09:24.895 "state": "online", 00:09:24.895 "raid_level": "raid1", 00:09:24.895 "superblock": true, 00:09:24.895 "num_base_bdevs": 3, 00:09:24.895 "num_base_bdevs_discovered": 3, 00:09:24.895 "num_base_bdevs_operational": 3, 00:09:24.895 "base_bdevs_list": [ 00:09:24.895 { 00:09:24.895 "name": "NewBaseBdev", 00:09:24.895 "uuid": "84da1d46-e0a9-4138-9346-49319a867567", 00:09:24.895 "is_configured": true, 00:09:24.895 "data_offset": 2048, 00:09:24.895 "data_size": 63488 00:09:24.895 }, 00:09:24.895 { 00:09:24.895 "name": "BaseBdev2", 00:09:24.895 "uuid": "4eb7842e-4f5d-4254-a806-8eb513340583", 00:09:24.895 "is_configured": true, 00:09:24.895 "data_offset": 2048, 00:09:24.895 "data_size": 63488 00:09:24.895 }, 00:09:24.895 { 00:09:24.895 "name": "BaseBdev3", 00:09:24.895 "uuid": "2aded048-ac16-417d-aa08-7719100fd889", 00:09:24.895 "is_configured": true, 00:09:24.895 "data_offset": 2048, 00:09:24.895 "data_size": 63488 00:09:24.895 } 00:09:24.895 ] 00:09:24.895 } 00:09:24.895 } 00:09:24.895 }' 00:09:24.895 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:24.895 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:24.895 BaseBdev2 00:09:24.895 BaseBdev3' 00:09:24.895 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:24.895 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:24.895 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:24.895 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:24.895 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:24.895 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:24.895 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:24.895 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:24.895 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:24.895 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:24.895 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:24.895 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:24.895 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:24.895 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:24.895 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:24.895 15:13:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:24.895 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:24.895 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:24.895 15:13:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:25.155 15:13:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:25.155 15:13:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:25.155 15:13:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.155 15:13:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:25.155 15:13:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.155 15:13:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:25.155 15:13:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:25.155 15:13:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:25.155 15:13:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.155 15:13:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:25.155 [2024-11-27 15:13:53.056575] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:25.155 [2024-11-27 15:13:53.056622] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:25.155 [2024-11-27 15:13:53.056717] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:25.155 [2024-11-27 15:13:53.056977] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:25.155 [2024-11-27 15:13:53.056989] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name Existed_Raid, state offline 00:09:25.155 15:13:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.155 15:13:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 79224 00:09:25.155 15:13:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 79224 ']' 00:09:25.155 15:13:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 79224 00:09:25.155 15:13:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:09:25.155 15:13:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:25.155 15:13:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 79224 00:09:25.155 15:13:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:25.155 15:13:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:25.155 15:13:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 79224' 00:09:25.155 killing process with pid 79224 00:09:25.155 15:13:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 79224 00:09:25.155 [2024-11-27 15:13:53.105656] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:25.155 15:13:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 79224 00:09:25.155 [2024-11-27 15:13:53.137783] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:25.416 15:13:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:09:25.416 ************************************ 00:09:25.416 END TEST raid_state_function_test_sb 00:09:25.416 ************************************ 00:09:25.416 00:09:25.416 real 0m8.719s 00:09:25.416 user 0m14.849s 00:09:25.416 sys 0m1.830s 00:09:25.416 15:13:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:25.416 15:13:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:25.416 15:13:53 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 3 00:09:25.416 15:13:53 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:09:25.416 15:13:53 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:25.416 15:13:53 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:25.416 ************************************ 00:09:25.416 START TEST raid_superblock_test 00:09:25.416 ************************************ 00:09:25.416 15:13:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 3 00:09:25.416 15:13:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:09:25.416 15:13:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:09:25.416 15:13:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:09:25.416 15:13:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:09:25.416 15:13:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:09:25.416 15:13:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:09:25.416 15:13:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:09:25.416 15:13:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:09:25.416 15:13:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:09:25.416 15:13:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:09:25.416 15:13:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:09:25.416 15:13:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:09:25.416 15:13:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:09:25.416 15:13:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:09:25.416 15:13:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:09:25.416 15:13:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=79822 00:09:25.416 15:13:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 79822 00:09:25.416 15:13:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:09:25.416 15:13:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 79822 ']' 00:09:25.416 15:13:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:25.416 15:13:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:25.416 15:13:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:25.416 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:25.416 15:13:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:25.416 15:13:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.676 [2024-11-27 15:13:53.524023] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:09:25.676 [2024-11-27 15:13:53.524289] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid79822 ] 00:09:25.676 [2024-11-27 15:13:53.702833] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:25.676 [2024-11-27 15:13:53.732616] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:25.676 [2024-11-27 15:13:53.775701] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:25.676 [2024-11-27 15:13:53.775817] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.615 malloc1 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.615 [2024-11-27 15:13:54.375960] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:26.615 [2024-11-27 15:13:54.376098] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:26.615 [2024-11-27 15:13:54.376136] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:09:26.615 [2024-11-27 15:13:54.376192] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:26.615 [2024-11-27 15:13:54.378303] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:26.615 [2024-11-27 15:13:54.378374] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:26.615 pt1 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.615 malloc2 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.615 [2024-11-27 15:13:54.408884] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:26.615 [2024-11-27 15:13:54.409042] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:26.615 [2024-11-27 15:13:54.409088] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:09:26.615 [2024-11-27 15:13:54.409120] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:26.615 [2024-11-27 15:13:54.411277] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:26.615 [2024-11-27 15:13:54.411350] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:26.615 pt2 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.615 malloc3 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.615 [2024-11-27 15:13:54.437946] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:26.615 [2024-11-27 15:13:54.438082] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:26.615 [2024-11-27 15:13:54.438122] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:09:26.615 [2024-11-27 15:13:54.438152] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:26.615 [2024-11-27 15:13:54.440314] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:26.615 [2024-11-27 15:13:54.440400] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:26.615 pt3 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.615 [2024-11-27 15:13:54.449984] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:26.615 [2024-11-27 15:13:54.451982] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:26.615 [2024-11-27 15:13:54.452088] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:26.615 [2024-11-27 15:13:54.452277] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:09:26.615 [2024-11-27 15:13:54.452330] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:26.615 [2024-11-27 15:13:54.452675] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:09:26.615 [2024-11-27 15:13:54.452885] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:09:26.615 [2024-11-27 15:13:54.452940] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:09:26.615 [2024-11-27 15:13:54.453160] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:26.615 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:26.616 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:26.616 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:26.616 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:26.616 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:26.616 15:13:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.616 15:13:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.616 15:13:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.616 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:26.616 "name": "raid_bdev1", 00:09:26.616 "uuid": "f7bca14f-a9b4-41c8-8957-b57bc18564ae", 00:09:26.616 "strip_size_kb": 0, 00:09:26.616 "state": "online", 00:09:26.616 "raid_level": "raid1", 00:09:26.616 "superblock": true, 00:09:26.616 "num_base_bdevs": 3, 00:09:26.616 "num_base_bdevs_discovered": 3, 00:09:26.616 "num_base_bdevs_operational": 3, 00:09:26.616 "base_bdevs_list": [ 00:09:26.616 { 00:09:26.616 "name": "pt1", 00:09:26.616 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:26.616 "is_configured": true, 00:09:26.616 "data_offset": 2048, 00:09:26.616 "data_size": 63488 00:09:26.616 }, 00:09:26.616 { 00:09:26.616 "name": "pt2", 00:09:26.616 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:26.616 "is_configured": true, 00:09:26.616 "data_offset": 2048, 00:09:26.616 "data_size": 63488 00:09:26.616 }, 00:09:26.616 { 00:09:26.616 "name": "pt3", 00:09:26.616 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:26.616 "is_configured": true, 00:09:26.616 "data_offset": 2048, 00:09:26.616 "data_size": 63488 00:09:26.616 } 00:09:26.616 ] 00:09:26.616 }' 00:09:26.616 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:26.616 15:13:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.875 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:09:26.875 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:26.875 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:26.875 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:26.875 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:26.875 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:26.875 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:26.875 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:26.875 15:13:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.875 15:13:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.876 [2024-11-27 15:13:54.881449] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:26.876 15:13:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.876 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:26.876 "name": "raid_bdev1", 00:09:26.876 "aliases": [ 00:09:26.876 "f7bca14f-a9b4-41c8-8957-b57bc18564ae" 00:09:26.876 ], 00:09:26.876 "product_name": "Raid Volume", 00:09:26.876 "block_size": 512, 00:09:26.876 "num_blocks": 63488, 00:09:26.876 "uuid": "f7bca14f-a9b4-41c8-8957-b57bc18564ae", 00:09:26.876 "assigned_rate_limits": { 00:09:26.876 "rw_ios_per_sec": 0, 00:09:26.876 "rw_mbytes_per_sec": 0, 00:09:26.876 "r_mbytes_per_sec": 0, 00:09:26.876 "w_mbytes_per_sec": 0 00:09:26.876 }, 00:09:26.876 "claimed": false, 00:09:26.876 "zoned": false, 00:09:26.876 "supported_io_types": { 00:09:26.876 "read": true, 00:09:26.876 "write": true, 00:09:26.876 "unmap": false, 00:09:26.876 "flush": false, 00:09:26.876 "reset": true, 00:09:26.876 "nvme_admin": false, 00:09:26.876 "nvme_io": false, 00:09:26.876 "nvme_io_md": false, 00:09:26.876 "write_zeroes": true, 00:09:26.876 "zcopy": false, 00:09:26.876 "get_zone_info": false, 00:09:26.876 "zone_management": false, 00:09:26.876 "zone_append": false, 00:09:26.876 "compare": false, 00:09:26.876 "compare_and_write": false, 00:09:26.876 "abort": false, 00:09:26.876 "seek_hole": false, 00:09:26.876 "seek_data": false, 00:09:26.876 "copy": false, 00:09:26.876 "nvme_iov_md": false 00:09:26.876 }, 00:09:26.876 "memory_domains": [ 00:09:26.876 { 00:09:26.876 "dma_device_id": "system", 00:09:26.876 "dma_device_type": 1 00:09:26.876 }, 00:09:26.876 { 00:09:26.876 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:26.876 "dma_device_type": 2 00:09:26.876 }, 00:09:26.876 { 00:09:26.876 "dma_device_id": "system", 00:09:26.876 "dma_device_type": 1 00:09:26.876 }, 00:09:26.876 { 00:09:26.876 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:26.876 "dma_device_type": 2 00:09:26.876 }, 00:09:26.876 { 00:09:26.876 "dma_device_id": "system", 00:09:26.876 "dma_device_type": 1 00:09:26.876 }, 00:09:26.876 { 00:09:26.876 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:26.876 "dma_device_type": 2 00:09:26.876 } 00:09:26.876 ], 00:09:26.876 "driver_specific": { 00:09:26.876 "raid": { 00:09:26.876 "uuid": "f7bca14f-a9b4-41c8-8957-b57bc18564ae", 00:09:26.876 "strip_size_kb": 0, 00:09:26.876 "state": "online", 00:09:26.876 "raid_level": "raid1", 00:09:26.876 "superblock": true, 00:09:26.876 "num_base_bdevs": 3, 00:09:26.876 "num_base_bdevs_discovered": 3, 00:09:26.876 "num_base_bdevs_operational": 3, 00:09:26.876 "base_bdevs_list": [ 00:09:26.876 { 00:09:26.876 "name": "pt1", 00:09:26.876 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:26.876 "is_configured": true, 00:09:26.876 "data_offset": 2048, 00:09:26.876 "data_size": 63488 00:09:26.876 }, 00:09:26.876 { 00:09:26.876 "name": "pt2", 00:09:26.876 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:26.876 "is_configured": true, 00:09:26.876 "data_offset": 2048, 00:09:26.876 "data_size": 63488 00:09:26.876 }, 00:09:26.876 { 00:09:26.876 "name": "pt3", 00:09:26.876 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:26.876 "is_configured": true, 00:09:26.876 "data_offset": 2048, 00:09:26.876 "data_size": 63488 00:09:26.876 } 00:09:26.876 ] 00:09:26.876 } 00:09:26.876 } 00:09:26.876 }' 00:09:26.876 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:26.876 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:26.876 pt2 00:09:26.876 pt3' 00:09:26.876 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:27.136 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:27.136 15:13:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.136 [2024-11-27 15:13:55.156992] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=f7bca14f-a9b4-41c8-8957-b57bc18564ae 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z f7bca14f-a9b4-41c8-8957-b57bc18564ae ']' 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.136 [2024-11-27 15:13:55.188628] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:27.136 [2024-11-27 15:13:55.188723] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:27.136 [2024-11-27 15:13:55.188831] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:27.136 [2024-11-27 15:13:55.188937] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:27.136 [2024-11-27 15:13:55.188990] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.136 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.396 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:09:27.396 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:09:27.396 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:27.396 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:09:27.396 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.396 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.396 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.396 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:27.396 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:09:27.396 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.396 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.396 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.396 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:27.396 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:09:27.396 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.396 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.396 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.396 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:09:27.396 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.396 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.396 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:09:27.396 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.396 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:09:27.396 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:27.396 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:09:27.396 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:27.396 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:09:27.396 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:27.396 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:09:27.396 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:27.396 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.397 [2024-11-27 15:13:55.344311] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:09:27.397 [2024-11-27 15:13:55.346158] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:09:27.397 [2024-11-27 15:13:55.346245] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:09:27.397 [2024-11-27 15:13:55.346319] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:09:27.397 [2024-11-27 15:13:55.346399] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:09:27.397 [2024-11-27 15:13:55.346467] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:09:27.397 [2024-11-27 15:13:55.346517] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:27.397 [2024-11-27 15:13:55.346547] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state configuring 00:09:27.397 request: 00:09:27.397 { 00:09:27.397 "name": "raid_bdev1", 00:09:27.397 "raid_level": "raid1", 00:09:27.397 "base_bdevs": [ 00:09:27.397 "malloc1", 00:09:27.397 "malloc2", 00:09:27.397 "malloc3" 00:09:27.397 ], 00:09:27.397 "superblock": false, 00:09:27.397 "method": "bdev_raid_create", 00:09:27.397 "req_id": 1 00:09:27.397 } 00:09:27.397 Got JSON-RPC error response 00:09:27.397 response: 00:09:27.397 { 00:09:27.397 "code": -17, 00:09:27.397 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:09:27.397 } 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.397 [2024-11-27 15:13:55.408191] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:27.397 [2024-11-27 15:13:55.408282] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:27.397 [2024-11-27 15:13:55.408320] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:09:27.397 [2024-11-27 15:13:55.408353] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:27.397 [2024-11-27 15:13:55.410487] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:27.397 [2024-11-27 15:13:55.410556] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:27.397 [2024-11-27 15:13:55.410655] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:09:27.397 [2024-11-27 15:13:55.410708] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:27.397 pt1 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:27.397 "name": "raid_bdev1", 00:09:27.397 "uuid": "f7bca14f-a9b4-41c8-8957-b57bc18564ae", 00:09:27.397 "strip_size_kb": 0, 00:09:27.397 "state": "configuring", 00:09:27.397 "raid_level": "raid1", 00:09:27.397 "superblock": true, 00:09:27.397 "num_base_bdevs": 3, 00:09:27.397 "num_base_bdevs_discovered": 1, 00:09:27.397 "num_base_bdevs_operational": 3, 00:09:27.397 "base_bdevs_list": [ 00:09:27.397 { 00:09:27.397 "name": "pt1", 00:09:27.397 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:27.397 "is_configured": true, 00:09:27.397 "data_offset": 2048, 00:09:27.397 "data_size": 63488 00:09:27.397 }, 00:09:27.397 { 00:09:27.397 "name": null, 00:09:27.397 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:27.397 "is_configured": false, 00:09:27.397 "data_offset": 2048, 00:09:27.397 "data_size": 63488 00:09:27.397 }, 00:09:27.397 { 00:09:27.397 "name": null, 00:09:27.397 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:27.397 "is_configured": false, 00:09:27.397 "data_offset": 2048, 00:09:27.397 "data_size": 63488 00:09:27.397 } 00:09:27.397 ] 00:09:27.397 }' 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:27.397 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.965 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:09:27.965 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:27.965 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.965 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.965 [2024-11-27 15:13:55.795524] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:27.965 [2024-11-27 15:13:55.795626] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:27.965 [2024-11-27 15:13:55.795662] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:09:27.965 [2024-11-27 15:13:55.795693] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:27.965 [2024-11-27 15:13:55.796134] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:27.965 [2024-11-27 15:13:55.796193] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:27.965 [2024-11-27 15:13:55.796292] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:27.965 [2024-11-27 15:13:55.796346] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:27.965 pt2 00:09:27.965 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.965 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:09:27.965 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.965 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.965 [2024-11-27 15:13:55.807523] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:09:27.965 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.965 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:09:27.965 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:27.965 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:27.965 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:27.966 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:27.966 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:27.966 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:27.966 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:27.966 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:27.966 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:27.966 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:27.966 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:27.966 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.966 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.966 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.966 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:27.966 "name": "raid_bdev1", 00:09:27.966 "uuid": "f7bca14f-a9b4-41c8-8957-b57bc18564ae", 00:09:27.966 "strip_size_kb": 0, 00:09:27.966 "state": "configuring", 00:09:27.966 "raid_level": "raid1", 00:09:27.966 "superblock": true, 00:09:27.966 "num_base_bdevs": 3, 00:09:27.966 "num_base_bdevs_discovered": 1, 00:09:27.966 "num_base_bdevs_operational": 3, 00:09:27.966 "base_bdevs_list": [ 00:09:27.966 { 00:09:27.966 "name": "pt1", 00:09:27.966 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:27.966 "is_configured": true, 00:09:27.966 "data_offset": 2048, 00:09:27.966 "data_size": 63488 00:09:27.966 }, 00:09:27.966 { 00:09:27.966 "name": null, 00:09:27.966 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:27.966 "is_configured": false, 00:09:27.966 "data_offset": 0, 00:09:27.966 "data_size": 63488 00:09:27.966 }, 00:09:27.966 { 00:09:27.966 "name": null, 00:09:27.966 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:27.966 "is_configured": false, 00:09:27.966 "data_offset": 2048, 00:09:27.966 "data_size": 63488 00:09:27.966 } 00:09:27.966 ] 00:09:27.966 }' 00:09:27.966 15:13:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:27.966 15:13:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.225 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:09:28.225 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:28.225 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:28.225 15:13:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.225 15:13:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.225 [2024-11-27 15:13:56.242754] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:28.225 [2024-11-27 15:13:56.242857] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:28.225 [2024-11-27 15:13:56.242895] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:09:28.225 [2024-11-27 15:13:56.242932] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:28.225 [2024-11-27 15:13:56.243305] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:28.225 [2024-11-27 15:13:56.243363] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:28.225 [2024-11-27 15:13:56.243458] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:28.225 [2024-11-27 15:13:56.243504] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:28.225 pt2 00:09:28.225 15:13:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.225 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:28.225 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:28.225 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:28.225 15:13:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.225 15:13:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.225 [2024-11-27 15:13:56.254718] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:28.226 [2024-11-27 15:13:56.254792] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:28.226 [2024-11-27 15:13:56.254828] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:09:28.226 [2024-11-27 15:13:56.254858] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:28.226 [2024-11-27 15:13:56.255171] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:28.226 [2024-11-27 15:13:56.255244] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:28.226 [2024-11-27 15:13:56.255329] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:09:28.226 [2024-11-27 15:13:56.255382] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:28.226 [2024-11-27 15:13:56.255502] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:09:28.226 [2024-11-27 15:13:56.255539] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:28.226 [2024-11-27 15:13:56.255783] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:09:28.226 [2024-11-27 15:13:56.255942] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:09:28.226 [2024-11-27 15:13:56.255985] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:09:28.226 [2024-11-27 15:13:56.256113] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:28.226 pt3 00:09:28.226 15:13:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.226 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:28.226 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:28.226 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:28.226 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:28.226 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:28.226 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:28.226 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:28.226 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:28.226 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:28.226 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:28.226 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:28.226 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:28.226 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:28.226 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:28.226 15:13:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.226 15:13:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.226 15:13:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.226 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:28.226 "name": "raid_bdev1", 00:09:28.226 "uuid": "f7bca14f-a9b4-41c8-8957-b57bc18564ae", 00:09:28.226 "strip_size_kb": 0, 00:09:28.226 "state": "online", 00:09:28.226 "raid_level": "raid1", 00:09:28.226 "superblock": true, 00:09:28.226 "num_base_bdevs": 3, 00:09:28.226 "num_base_bdevs_discovered": 3, 00:09:28.226 "num_base_bdevs_operational": 3, 00:09:28.226 "base_bdevs_list": [ 00:09:28.226 { 00:09:28.226 "name": "pt1", 00:09:28.226 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:28.226 "is_configured": true, 00:09:28.226 "data_offset": 2048, 00:09:28.226 "data_size": 63488 00:09:28.226 }, 00:09:28.226 { 00:09:28.226 "name": "pt2", 00:09:28.226 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:28.226 "is_configured": true, 00:09:28.226 "data_offset": 2048, 00:09:28.226 "data_size": 63488 00:09:28.226 }, 00:09:28.226 { 00:09:28.226 "name": "pt3", 00:09:28.226 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:28.226 "is_configured": true, 00:09:28.226 "data_offset": 2048, 00:09:28.226 "data_size": 63488 00:09:28.226 } 00:09:28.226 ] 00:09:28.226 }' 00:09:28.226 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:28.226 15:13:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.793 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:09:28.793 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:28.793 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:28.793 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:28.793 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:28.793 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:28.793 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:28.793 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:28.793 15:13:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.793 15:13:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.793 [2024-11-27 15:13:56.722224] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:28.793 15:13:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.793 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:28.793 "name": "raid_bdev1", 00:09:28.793 "aliases": [ 00:09:28.793 "f7bca14f-a9b4-41c8-8957-b57bc18564ae" 00:09:28.793 ], 00:09:28.793 "product_name": "Raid Volume", 00:09:28.793 "block_size": 512, 00:09:28.793 "num_blocks": 63488, 00:09:28.793 "uuid": "f7bca14f-a9b4-41c8-8957-b57bc18564ae", 00:09:28.793 "assigned_rate_limits": { 00:09:28.793 "rw_ios_per_sec": 0, 00:09:28.793 "rw_mbytes_per_sec": 0, 00:09:28.793 "r_mbytes_per_sec": 0, 00:09:28.793 "w_mbytes_per_sec": 0 00:09:28.793 }, 00:09:28.793 "claimed": false, 00:09:28.793 "zoned": false, 00:09:28.793 "supported_io_types": { 00:09:28.793 "read": true, 00:09:28.793 "write": true, 00:09:28.793 "unmap": false, 00:09:28.793 "flush": false, 00:09:28.793 "reset": true, 00:09:28.793 "nvme_admin": false, 00:09:28.793 "nvme_io": false, 00:09:28.793 "nvme_io_md": false, 00:09:28.793 "write_zeroes": true, 00:09:28.793 "zcopy": false, 00:09:28.793 "get_zone_info": false, 00:09:28.793 "zone_management": false, 00:09:28.793 "zone_append": false, 00:09:28.793 "compare": false, 00:09:28.793 "compare_and_write": false, 00:09:28.793 "abort": false, 00:09:28.793 "seek_hole": false, 00:09:28.793 "seek_data": false, 00:09:28.793 "copy": false, 00:09:28.793 "nvme_iov_md": false 00:09:28.793 }, 00:09:28.793 "memory_domains": [ 00:09:28.793 { 00:09:28.793 "dma_device_id": "system", 00:09:28.793 "dma_device_type": 1 00:09:28.793 }, 00:09:28.793 { 00:09:28.793 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:28.793 "dma_device_type": 2 00:09:28.793 }, 00:09:28.793 { 00:09:28.793 "dma_device_id": "system", 00:09:28.793 "dma_device_type": 1 00:09:28.793 }, 00:09:28.793 { 00:09:28.793 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:28.793 "dma_device_type": 2 00:09:28.793 }, 00:09:28.793 { 00:09:28.793 "dma_device_id": "system", 00:09:28.793 "dma_device_type": 1 00:09:28.793 }, 00:09:28.793 { 00:09:28.793 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:28.793 "dma_device_type": 2 00:09:28.793 } 00:09:28.793 ], 00:09:28.793 "driver_specific": { 00:09:28.793 "raid": { 00:09:28.793 "uuid": "f7bca14f-a9b4-41c8-8957-b57bc18564ae", 00:09:28.793 "strip_size_kb": 0, 00:09:28.793 "state": "online", 00:09:28.793 "raid_level": "raid1", 00:09:28.793 "superblock": true, 00:09:28.793 "num_base_bdevs": 3, 00:09:28.793 "num_base_bdevs_discovered": 3, 00:09:28.793 "num_base_bdevs_operational": 3, 00:09:28.793 "base_bdevs_list": [ 00:09:28.793 { 00:09:28.793 "name": "pt1", 00:09:28.793 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:28.793 "is_configured": true, 00:09:28.793 "data_offset": 2048, 00:09:28.793 "data_size": 63488 00:09:28.793 }, 00:09:28.793 { 00:09:28.793 "name": "pt2", 00:09:28.793 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:28.793 "is_configured": true, 00:09:28.793 "data_offset": 2048, 00:09:28.793 "data_size": 63488 00:09:28.793 }, 00:09:28.793 { 00:09:28.793 "name": "pt3", 00:09:28.793 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:28.793 "is_configured": true, 00:09:28.793 "data_offset": 2048, 00:09:28.793 "data_size": 63488 00:09:28.793 } 00:09:28.793 ] 00:09:28.793 } 00:09:28.793 } 00:09:28.793 }' 00:09:28.793 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:28.793 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:28.793 pt2 00:09:28.793 pt3' 00:09:28.793 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:28.793 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:28.793 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:28.793 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:28.793 15:13:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.793 15:13:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.793 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:28.793 15:13:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.793 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:28.793 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:28.793 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:28.793 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:28.793 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:28.793 15:13:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.793 15:13:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.053 15:13:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.053 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:29.053 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:29.053 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:29.053 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:29.053 15:13:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.053 15:13:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.053 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:29.053 15:13:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.053 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:29.053 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:29.053 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:09:29.053 15:13:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:29.053 15:13:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.053 15:13:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.053 [2024-11-27 15:13:57.001673] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:29.053 15:13:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.053 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' f7bca14f-a9b4-41c8-8957-b57bc18564ae '!=' f7bca14f-a9b4-41c8-8957-b57bc18564ae ']' 00:09:29.053 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:09:29.053 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:29.053 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:29.053 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:09:29.053 15:13:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.053 15:13:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.053 [2024-11-27 15:13:57.049377] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:09:29.053 15:13:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.053 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:29.053 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:29.053 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:29.053 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:29.053 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:29.053 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:29.053 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:29.053 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:29.053 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:29.053 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:29.053 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:29.053 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:29.053 15:13:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.053 15:13:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.053 15:13:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.053 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:29.053 "name": "raid_bdev1", 00:09:29.053 "uuid": "f7bca14f-a9b4-41c8-8957-b57bc18564ae", 00:09:29.053 "strip_size_kb": 0, 00:09:29.053 "state": "online", 00:09:29.053 "raid_level": "raid1", 00:09:29.053 "superblock": true, 00:09:29.053 "num_base_bdevs": 3, 00:09:29.053 "num_base_bdevs_discovered": 2, 00:09:29.053 "num_base_bdevs_operational": 2, 00:09:29.053 "base_bdevs_list": [ 00:09:29.053 { 00:09:29.053 "name": null, 00:09:29.053 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:29.053 "is_configured": false, 00:09:29.053 "data_offset": 0, 00:09:29.053 "data_size": 63488 00:09:29.053 }, 00:09:29.053 { 00:09:29.053 "name": "pt2", 00:09:29.053 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:29.053 "is_configured": true, 00:09:29.053 "data_offset": 2048, 00:09:29.053 "data_size": 63488 00:09:29.053 }, 00:09:29.053 { 00:09:29.053 "name": "pt3", 00:09:29.053 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:29.053 "is_configured": true, 00:09:29.053 "data_offset": 2048, 00:09:29.053 "data_size": 63488 00:09:29.053 } 00:09:29.053 ] 00:09:29.053 }' 00:09:29.053 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:29.053 15:13:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.622 [2024-11-27 15:13:57.528517] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:29.622 [2024-11-27 15:13:57.528583] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:29.622 [2024-11-27 15:13:57.528657] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:29.622 [2024-11-27 15:13:57.528726] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:29.622 [2024-11-27 15:13:57.528790] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.622 [2024-11-27 15:13:57.612385] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:29.622 [2024-11-27 15:13:57.612467] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:29.622 [2024-11-27 15:13:57.612500] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:09:29.622 [2024-11-27 15:13:57.612527] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:29.622 [2024-11-27 15:13:57.614589] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:29.622 [2024-11-27 15:13:57.614669] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:29.622 [2024-11-27 15:13:57.614756] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:29.622 [2024-11-27 15:13:57.614807] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:29.622 pt2 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.622 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:29.622 "name": "raid_bdev1", 00:09:29.622 "uuid": "f7bca14f-a9b4-41c8-8957-b57bc18564ae", 00:09:29.622 "strip_size_kb": 0, 00:09:29.622 "state": "configuring", 00:09:29.622 "raid_level": "raid1", 00:09:29.622 "superblock": true, 00:09:29.623 "num_base_bdevs": 3, 00:09:29.623 "num_base_bdevs_discovered": 1, 00:09:29.623 "num_base_bdevs_operational": 2, 00:09:29.623 "base_bdevs_list": [ 00:09:29.623 { 00:09:29.623 "name": null, 00:09:29.623 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:29.623 "is_configured": false, 00:09:29.623 "data_offset": 2048, 00:09:29.623 "data_size": 63488 00:09:29.623 }, 00:09:29.623 { 00:09:29.623 "name": "pt2", 00:09:29.623 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:29.623 "is_configured": true, 00:09:29.623 "data_offset": 2048, 00:09:29.623 "data_size": 63488 00:09:29.623 }, 00:09:29.623 { 00:09:29.623 "name": null, 00:09:29.623 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:29.623 "is_configured": false, 00:09:29.623 "data_offset": 2048, 00:09:29.623 "data_size": 63488 00:09:29.623 } 00:09:29.623 ] 00:09:29.623 }' 00:09:29.623 15:13:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:29.623 15:13:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.191 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:09:30.191 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:09:30.191 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=2 00:09:30.191 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:30.191 15:13:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.191 15:13:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.191 [2024-11-27 15:13:58.039696] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:30.191 [2024-11-27 15:13:58.039785] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:30.191 [2024-11-27 15:13:58.039824] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:09:30.191 [2024-11-27 15:13:58.039852] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:30.191 [2024-11-27 15:13:58.040248] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:30.191 [2024-11-27 15:13:58.040310] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:30.192 [2024-11-27 15:13:58.040410] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:09:30.192 [2024-11-27 15:13:58.040459] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:30.192 [2024-11-27 15:13:58.040567] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:09:30.192 [2024-11-27 15:13:58.040604] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:30.192 [2024-11-27 15:13:58.040870] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:09:30.192 [2024-11-27 15:13:58.041064] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:09:30.192 [2024-11-27 15:13:58.041083] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006d00 00:09:30.192 [2024-11-27 15:13:58.041183] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:30.192 pt3 00:09:30.192 15:13:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.192 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:30.192 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:30.192 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:30.192 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:30.192 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:30.192 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:30.192 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:30.192 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:30.192 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:30.192 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:30.192 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:30.192 15:13:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.192 15:13:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.192 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:30.192 15:13:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.192 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:30.192 "name": "raid_bdev1", 00:09:30.192 "uuid": "f7bca14f-a9b4-41c8-8957-b57bc18564ae", 00:09:30.192 "strip_size_kb": 0, 00:09:30.192 "state": "online", 00:09:30.192 "raid_level": "raid1", 00:09:30.192 "superblock": true, 00:09:30.192 "num_base_bdevs": 3, 00:09:30.192 "num_base_bdevs_discovered": 2, 00:09:30.192 "num_base_bdevs_operational": 2, 00:09:30.192 "base_bdevs_list": [ 00:09:30.192 { 00:09:30.192 "name": null, 00:09:30.192 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:30.192 "is_configured": false, 00:09:30.192 "data_offset": 2048, 00:09:30.192 "data_size": 63488 00:09:30.192 }, 00:09:30.192 { 00:09:30.192 "name": "pt2", 00:09:30.192 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:30.192 "is_configured": true, 00:09:30.192 "data_offset": 2048, 00:09:30.192 "data_size": 63488 00:09:30.192 }, 00:09:30.192 { 00:09:30.192 "name": "pt3", 00:09:30.192 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:30.192 "is_configured": true, 00:09:30.192 "data_offset": 2048, 00:09:30.192 "data_size": 63488 00:09:30.192 } 00:09:30.192 ] 00:09:30.192 }' 00:09:30.192 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:30.192 15:13:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.452 [2024-11-27 15:13:58.470877] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:30.452 [2024-11-27 15:13:58.470952] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:30.452 [2024-11-27 15:13:58.471032] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:30.452 [2024-11-27 15:13:58.471102] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:30.452 [2024-11-27 15:13:58.471136] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name raid_bdev1, state offline 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 3 -gt 2 ']' 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@534 -- # i=2 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt3 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.452 [2024-11-27 15:13:58.538765] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:30.452 [2024-11-27 15:13:58.538851] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:30.452 [2024-11-27 15:13:58.538883] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:09:30.452 [2024-11-27 15:13:58.538923] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:30.452 [2024-11-27 15:13:58.540997] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:30.452 [2024-11-27 15:13:58.541078] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:30.452 [2024-11-27 15:13:58.541160] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:09:30.452 [2024-11-27 15:13:58.541216] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:30.452 [2024-11-27 15:13:58.541331] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:09:30.452 [2024-11-27 15:13:58.541386] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:30.452 [2024-11-27 15:13:58.541431] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007080 name raid_bdev1, state configuring 00:09:30.452 [2024-11-27 15:13:58.541524] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:30.452 pt1 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 3 -gt 2 ']' 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.452 15:13:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.711 15:13:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.711 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:30.711 "name": "raid_bdev1", 00:09:30.711 "uuid": "f7bca14f-a9b4-41c8-8957-b57bc18564ae", 00:09:30.711 "strip_size_kb": 0, 00:09:30.711 "state": "configuring", 00:09:30.711 "raid_level": "raid1", 00:09:30.711 "superblock": true, 00:09:30.711 "num_base_bdevs": 3, 00:09:30.711 "num_base_bdevs_discovered": 1, 00:09:30.711 "num_base_bdevs_operational": 2, 00:09:30.711 "base_bdevs_list": [ 00:09:30.711 { 00:09:30.711 "name": null, 00:09:30.711 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:30.711 "is_configured": false, 00:09:30.711 "data_offset": 2048, 00:09:30.711 "data_size": 63488 00:09:30.711 }, 00:09:30.711 { 00:09:30.711 "name": "pt2", 00:09:30.711 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:30.711 "is_configured": true, 00:09:30.711 "data_offset": 2048, 00:09:30.711 "data_size": 63488 00:09:30.711 }, 00:09:30.711 { 00:09:30.711 "name": null, 00:09:30.711 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:30.711 "is_configured": false, 00:09:30.711 "data_offset": 2048, 00:09:30.711 "data_size": 63488 00:09:30.711 } 00:09:30.711 ] 00:09:30.711 }' 00:09:30.711 15:13:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:30.711 15:13:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.972 15:13:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:09:30.972 15:13:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:09:30.972 15:13:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.972 15:13:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.972 15:13:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.972 15:13:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:09:30.972 15:13:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:30.972 15:13:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.972 15:13:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.972 [2024-11-27 15:13:59.053876] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:30.972 [2024-11-27 15:13:59.053992] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:30.972 [2024-11-27 15:13:59.054031] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:09:30.972 [2024-11-27 15:13:59.054061] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:30.972 [2024-11-27 15:13:59.054460] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:30.972 [2024-11-27 15:13:59.054519] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:30.972 [2024-11-27 15:13:59.054612] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:09:30.972 [2024-11-27 15:13:59.054664] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:30.972 [2024-11-27 15:13:59.054775] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007400 00:09:30.972 [2024-11-27 15:13:59.054816] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:30.972 [2024-11-27 15:13:59.055083] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:09:30.972 [2024-11-27 15:13:59.055269] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007400 00:09:30.972 [2024-11-27 15:13:59.055309] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007400 00:09:30.972 [2024-11-27 15:13:59.055445] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:30.972 pt3 00:09:30.972 15:13:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.972 15:13:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:30.972 15:13:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:30.972 15:13:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:30.972 15:13:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:30.972 15:13:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:30.972 15:13:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:30.972 15:13:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:30.972 15:13:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:30.972 15:13:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:30.972 15:13:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:30.972 15:13:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:30.972 15:13:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:30.972 15:13:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.972 15:13:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.278 15:13:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.279 15:13:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:31.279 "name": "raid_bdev1", 00:09:31.279 "uuid": "f7bca14f-a9b4-41c8-8957-b57bc18564ae", 00:09:31.279 "strip_size_kb": 0, 00:09:31.279 "state": "online", 00:09:31.279 "raid_level": "raid1", 00:09:31.279 "superblock": true, 00:09:31.279 "num_base_bdevs": 3, 00:09:31.279 "num_base_bdevs_discovered": 2, 00:09:31.279 "num_base_bdevs_operational": 2, 00:09:31.279 "base_bdevs_list": [ 00:09:31.279 { 00:09:31.279 "name": null, 00:09:31.279 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:31.279 "is_configured": false, 00:09:31.279 "data_offset": 2048, 00:09:31.279 "data_size": 63488 00:09:31.279 }, 00:09:31.279 { 00:09:31.279 "name": "pt2", 00:09:31.279 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:31.279 "is_configured": true, 00:09:31.279 "data_offset": 2048, 00:09:31.279 "data_size": 63488 00:09:31.279 }, 00:09:31.279 { 00:09:31.279 "name": "pt3", 00:09:31.279 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:31.279 "is_configured": true, 00:09:31.279 "data_offset": 2048, 00:09:31.279 "data_size": 63488 00:09:31.279 } 00:09:31.279 ] 00:09:31.279 }' 00:09:31.279 15:13:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:31.279 15:13:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.578 15:13:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:09:31.578 15:13:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.578 15:13:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:09:31.578 15:13:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.578 15:13:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.578 15:13:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:09:31.578 15:13:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:31.578 15:13:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.578 15:13:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.578 15:13:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:09:31.578 [2024-11-27 15:13:59.505341] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:31.578 15:13:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.578 15:13:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' f7bca14f-a9b4-41c8-8957-b57bc18564ae '!=' f7bca14f-a9b4-41c8-8957-b57bc18564ae ']' 00:09:31.578 15:13:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 79822 00:09:31.578 15:13:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 79822 ']' 00:09:31.578 15:13:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 79822 00:09:31.578 15:13:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:09:31.578 15:13:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:31.578 15:13:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 79822 00:09:31.578 15:13:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:31.578 15:13:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:31.578 15:13:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 79822' 00:09:31.578 killing process with pid 79822 00:09:31.578 15:13:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 79822 00:09:31.578 [2024-11-27 15:13:59.581854] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:31.578 [2024-11-27 15:13:59.581975] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:31.578 [2024-11-27 15:13:59.582060] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:31.578 [2024-11-27 15:13:59.582072] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007400 name raid_bdev1, state offline 00:09:31.578 15:13:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 79822 00:09:31.578 [2024-11-27 15:13:59.615415] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:31.845 15:13:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:09:31.845 00:09:31.845 real 0m6.407s 00:09:31.845 user 0m10.692s 00:09:31.845 sys 0m1.391s 00:09:31.845 15:13:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:31.845 15:13:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.845 ************************************ 00:09:31.845 END TEST raid_superblock_test 00:09:31.845 ************************************ 00:09:31.845 15:13:59 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 3 read 00:09:31.845 15:13:59 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:31.845 15:13:59 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:31.845 15:13:59 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:31.845 ************************************ 00:09:31.845 START TEST raid_read_error_test 00:09:31.845 ************************************ 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 3 read 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.YVWNNUhtXK 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=80257 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 80257 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 80257 ']' 00:09:31.845 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:31.845 15:13:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:32.105 [2024-11-27 15:14:00.020866] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:09:32.105 [2024-11-27 15:14:00.021020] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid80257 ] 00:09:32.105 [2024-11-27 15:14:00.197855] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:32.365 [2024-11-27 15:14:00.225740] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:32.365 [2024-11-27 15:14:00.268822] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:32.365 [2024-11-27 15:14:00.268862] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:32.935 BaseBdev1_malloc 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:32.935 true 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:32.935 [2024-11-27 15:14:00.884513] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:32.935 [2024-11-27 15:14:00.884609] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:32.935 [2024-11-27 15:14:00.884640] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:09:32.935 [2024-11-27 15:14:00.884652] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:32.935 [2024-11-27 15:14:00.886740] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:32.935 [2024-11-27 15:14:00.886779] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:32.935 BaseBdev1 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:32.935 BaseBdev2_malloc 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:32.935 true 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:32.935 [2024-11-27 15:14:00.925119] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:32.935 [2024-11-27 15:14:00.925204] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:32.935 [2024-11-27 15:14:00.925254] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:09:32.935 [2024-11-27 15:14:00.925264] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:32.935 [2024-11-27 15:14:00.927292] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:32.935 [2024-11-27 15:14:00.927328] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:32.935 BaseBdev2 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:32.935 BaseBdev3_malloc 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:32.935 true 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:32.935 [2024-11-27 15:14:00.965626] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:32.935 [2024-11-27 15:14:00.965721] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:32.935 [2024-11-27 15:14:00.965773] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:09:32.935 [2024-11-27 15:14:00.965802] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:32.935 [2024-11-27 15:14:00.967834] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:32.935 [2024-11-27 15:14:00.967909] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:32.935 BaseBdev3 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:32.935 [2024-11-27 15:14:00.977652] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:32.935 [2024-11-27 15:14:00.979398] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:32.935 [2024-11-27 15:14:00.979507] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:32.935 [2024-11-27 15:14:00.979721] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:09:32.935 [2024-11-27 15:14:00.979774] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:32.935 [2024-11-27 15:14:00.980033] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:09:32.935 [2024-11-27 15:14:00.980207] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:09:32.935 [2024-11-27 15:14:00.980247] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006d00 00:09:32.935 [2024-11-27 15:14:00.980411] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:32.935 15:14:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:32.936 15:14:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:32.936 15:14:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:32.936 15:14:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:32.936 15:14:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:32.936 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.936 15:14:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:32.936 15:14:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.936 15:14:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:32.936 "name": "raid_bdev1", 00:09:32.936 "uuid": "b2feb67f-54f0-438a-bfc4-2c3a46ff1519", 00:09:32.936 "strip_size_kb": 0, 00:09:32.936 "state": "online", 00:09:32.936 "raid_level": "raid1", 00:09:32.936 "superblock": true, 00:09:32.936 "num_base_bdevs": 3, 00:09:32.936 "num_base_bdevs_discovered": 3, 00:09:32.936 "num_base_bdevs_operational": 3, 00:09:32.936 "base_bdevs_list": [ 00:09:32.936 { 00:09:32.936 "name": "BaseBdev1", 00:09:32.936 "uuid": "9ac1a9fb-012b-5766-a498-69d7591893b8", 00:09:32.936 "is_configured": true, 00:09:32.936 "data_offset": 2048, 00:09:32.936 "data_size": 63488 00:09:32.936 }, 00:09:32.936 { 00:09:32.936 "name": "BaseBdev2", 00:09:32.936 "uuid": "7f62c976-8a32-5295-84e9-45afd69269a1", 00:09:32.936 "is_configured": true, 00:09:32.936 "data_offset": 2048, 00:09:32.936 "data_size": 63488 00:09:32.936 }, 00:09:32.936 { 00:09:32.936 "name": "BaseBdev3", 00:09:32.936 "uuid": "35e2c993-71c2-5479-a4d0-cc1cac82c523", 00:09:32.936 "is_configured": true, 00:09:32.936 "data_offset": 2048, 00:09:32.936 "data_size": 63488 00:09:32.936 } 00:09:32.936 ] 00:09:32.936 }' 00:09:32.936 15:14:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:32.936 15:14:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.505 15:14:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:33.505 15:14:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:33.505 [2024-11-27 15:14:01.573105] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006560 00:09:34.443 15:14:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:09:34.443 15:14:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.443 15:14:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.443 15:14:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.443 15:14:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:34.443 15:14:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:09:34.443 15:14:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:09:34.443 15:14:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:09:34.443 15:14:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:34.443 15:14:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:34.443 15:14:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:34.443 15:14:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:34.443 15:14:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:34.443 15:14:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:34.443 15:14:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:34.443 15:14:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:34.443 15:14:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:34.443 15:14:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:34.443 15:14:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:34.443 15:14:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:34.443 15:14:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.443 15:14:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.443 15:14:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.703 15:14:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:34.703 "name": "raid_bdev1", 00:09:34.703 "uuid": "b2feb67f-54f0-438a-bfc4-2c3a46ff1519", 00:09:34.703 "strip_size_kb": 0, 00:09:34.703 "state": "online", 00:09:34.703 "raid_level": "raid1", 00:09:34.703 "superblock": true, 00:09:34.703 "num_base_bdevs": 3, 00:09:34.703 "num_base_bdevs_discovered": 3, 00:09:34.703 "num_base_bdevs_operational": 3, 00:09:34.703 "base_bdevs_list": [ 00:09:34.703 { 00:09:34.703 "name": "BaseBdev1", 00:09:34.703 "uuid": "9ac1a9fb-012b-5766-a498-69d7591893b8", 00:09:34.703 "is_configured": true, 00:09:34.703 "data_offset": 2048, 00:09:34.703 "data_size": 63488 00:09:34.703 }, 00:09:34.703 { 00:09:34.703 "name": "BaseBdev2", 00:09:34.703 "uuid": "7f62c976-8a32-5295-84e9-45afd69269a1", 00:09:34.703 "is_configured": true, 00:09:34.703 "data_offset": 2048, 00:09:34.703 "data_size": 63488 00:09:34.703 }, 00:09:34.703 { 00:09:34.703 "name": "BaseBdev3", 00:09:34.703 "uuid": "35e2c993-71c2-5479-a4d0-cc1cac82c523", 00:09:34.703 "is_configured": true, 00:09:34.703 "data_offset": 2048, 00:09:34.703 "data_size": 63488 00:09:34.703 } 00:09:34.703 ] 00:09:34.703 }' 00:09:34.703 15:14:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:34.703 15:14:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.963 15:14:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:34.963 15:14:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.963 15:14:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.963 [2024-11-27 15:14:02.976853] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:34.963 [2024-11-27 15:14:02.976985] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:34.963 [2024-11-27 15:14:02.979503] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:34.963 [2024-11-27 15:14:02.979585] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:34.963 [2024-11-27 15:14:02.979730] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:34.963 [2024-11-27 15:14:02.979807] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name raid_bdev1, state offline 00:09:34.963 { 00:09:34.963 "results": [ 00:09:34.963 { 00:09:34.963 "job": "raid_bdev1", 00:09:34.963 "core_mask": "0x1", 00:09:34.963 "workload": "randrw", 00:09:34.963 "percentage": 50, 00:09:34.963 "status": "finished", 00:09:34.963 "queue_depth": 1, 00:09:34.963 "io_size": 131072, 00:09:34.963 "runtime": 1.404929, 00:09:34.963 "iops": 14034.16115689832, 00:09:34.963 "mibps": 1754.27014461229, 00:09:34.963 "io_failed": 0, 00:09:34.963 "io_timeout": 0, 00:09:34.963 "avg_latency_us": 68.62897670154965, 00:09:34.963 "min_latency_us": 22.69344978165939, 00:09:34.963 "max_latency_us": 1430.9170305676855 00:09:34.963 } 00:09:34.963 ], 00:09:34.963 "core_count": 1 00:09:34.963 } 00:09:34.963 15:14:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.963 15:14:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 80257 00:09:34.963 15:14:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 80257 ']' 00:09:34.963 15:14:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 80257 00:09:34.963 15:14:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:09:34.963 15:14:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:34.963 15:14:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 80257 00:09:34.963 killing process with pid 80257 00:09:34.963 15:14:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:34.963 15:14:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:34.963 15:14:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 80257' 00:09:34.963 15:14:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 80257 00:09:34.963 [2024-11-27 15:14:03.018660] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:34.963 15:14:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 80257 00:09:34.963 [2024-11-27 15:14:03.044050] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:35.222 15:14:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.YVWNNUhtXK 00:09:35.222 15:14:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:35.222 15:14:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:35.222 15:14:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:09:35.222 15:14:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:09:35.222 15:14:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:35.222 15:14:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:35.222 15:14:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:09:35.222 ************************************ 00:09:35.222 END TEST raid_read_error_test 00:09:35.222 ************************************ 00:09:35.222 00:09:35.222 real 0m3.358s 00:09:35.222 user 0m4.333s 00:09:35.222 sys 0m0.538s 00:09:35.222 15:14:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:35.222 15:14:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.222 15:14:03 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 3 write 00:09:35.222 15:14:03 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:35.482 15:14:03 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:35.482 15:14:03 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:35.482 ************************************ 00:09:35.482 START TEST raid_write_error_test 00:09:35.482 ************************************ 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 3 write 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.qeNCYOfu8O 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=80391 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 80391 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 80391 ']' 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:35.482 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:35.482 15:14:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.482 [2024-11-27 15:14:03.447429] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:09:35.482 [2024-11-27 15:14:03.447639] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid80391 ] 00:09:35.742 [2024-11-27 15:14:03.612469] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:35.742 [2024-11-27 15:14:03.640413] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:35.742 [2024-11-27 15:14:03.683576] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:35.742 [2024-11-27 15:14:03.683646] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.312 BaseBdev1_malloc 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.312 true 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.312 [2024-11-27 15:14:04.315763] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:36.312 [2024-11-27 15:14:04.315869] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:36.312 [2024-11-27 15:14:04.315893] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:09:36.312 [2024-11-27 15:14:04.315922] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:36.312 [2024-11-27 15:14:04.318048] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:36.312 [2024-11-27 15:14:04.318132] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:36.312 BaseBdev1 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.312 BaseBdev2_malloc 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.312 true 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.312 [2024-11-27 15:14:04.356446] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:36.312 [2024-11-27 15:14:04.356536] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:36.312 [2024-11-27 15:14:04.356559] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:09:36.312 [2024-11-27 15:14:04.356568] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:36.312 [2024-11-27 15:14:04.358634] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:36.312 [2024-11-27 15:14:04.358673] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:36.312 BaseBdev2 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.312 BaseBdev3_malloc 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.312 true 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.312 [2024-11-27 15:14:04.397336] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:36.312 [2024-11-27 15:14:04.397428] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:36.312 [2024-11-27 15:14:04.397465] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:09:36.312 [2024-11-27 15:14:04.397493] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:36.312 [2024-11-27 15:14:04.399550] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:36.312 [2024-11-27 15:14:04.399631] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:36.312 BaseBdev3 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.312 [2024-11-27 15:14:04.409357] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:36.312 [2024-11-27 15:14:04.411189] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:36.312 [2024-11-27 15:14:04.411306] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:36.312 [2024-11-27 15:14:04.411494] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:09:36.312 [2024-11-27 15:14:04.411512] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:36.312 [2024-11-27 15:14:04.411787] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:09:36.312 [2024-11-27 15:14:04.411977] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:09:36.312 [2024-11-27 15:14:04.411995] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006d00 00:09:36.312 [2024-11-27 15:14:04.412141] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:36.312 15:14:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:36.572 15:14:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:36.572 15:14:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:36.572 15:14:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:36.572 15:14:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:36.572 15:14:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:36.572 15:14:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:36.572 15:14:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:36.572 15:14:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:36.572 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.572 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.572 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.572 15:14:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:36.572 "name": "raid_bdev1", 00:09:36.572 "uuid": "8b9288a3-9bd9-424c-9c9f-cdc14b7d2368", 00:09:36.572 "strip_size_kb": 0, 00:09:36.572 "state": "online", 00:09:36.572 "raid_level": "raid1", 00:09:36.572 "superblock": true, 00:09:36.572 "num_base_bdevs": 3, 00:09:36.572 "num_base_bdevs_discovered": 3, 00:09:36.572 "num_base_bdevs_operational": 3, 00:09:36.572 "base_bdevs_list": [ 00:09:36.572 { 00:09:36.572 "name": "BaseBdev1", 00:09:36.572 "uuid": "bd0a861d-f3da-56e6-9f16-158aa1540049", 00:09:36.572 "is_configured": true, 00:09:36.572 "data_offset": 2048, 00:09:36.572 "data_size": 63488 00:09:36.572 }, 00:09:36.572 { 00:09:36.572 "name": "BaseBdev2", 00:09:36.572 "uuid": "c31a1d78-bcbf-513c-86c0-e30dd5598d17", 00:09:36.572 "is_configured": true, 00:09:36.572 "data_offset": 2048, 00:09:36.572 "data_size": 63488 00:09:36.572 }, 00:09:36.572 { 00:09:36.572 "name": "BaseBdev3", 00:09:36.572 "uuid": "67d69fa3-51d6-5dc6-8ff0-fe245fe0209c", 00:09:36.572 "is_configured": true, 00:09:36.572 "data_offset": 2048, 00:09:36.572 "data_size": 63488 00:09:36.572 } 00:09:36.572 ] 00:09:36.572 }' 00:09:36.572 15:14:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:36.572 15:14:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.832 15:14:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:36.832 15:14:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:36.832 [2024-11-27 15:14:04.892931] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006560 00:09:37.772 15:14:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:09:37.772 15:14:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:37.772 15:14:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.772 [2024-11-27 15:14:05.809295] bdev_raid.c:2276:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:09:37.772 [2024-11-27 15:14:05.809442] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:37.772 [2024-11-27 15:14:05.809731] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000006560 00:09:37.772 15:14:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:37.772 15:14:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:37.772 15:14:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:09:37.772 15:14:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:09:37.772 15:14:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=2 00:09:37.772 15:14:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:37.772 15:14:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:37.772 15:14:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:37.772 15:14:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:37.772 15:14:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:37.772 15:14:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:37.772 15:14:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:37.772 15:14:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:37.772 15:14:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:37.772 15:14:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:37.772 15:14:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:37.772 15:14:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:37.772 15:14:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:37.772 15:14:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.772 15:14:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:37.772 15:14:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:37.772 "name": "raid_bdev1", 00:09:37.772 "uuid": "8b9288a3-9bd9-424c-9c9f-cdc14b7d2368", 00:09:37.772 "strip_size_kb": 0, 00:09:37.772 "state": "online", 00:09:37.772 "raid_level": "raid1", 00:09:37.772 "superblock": true, 00:09:37.772 "num_base_bdevs": 3, 00:09:37.772 "num_base_bdevs_discovered": 2, 00:09:37.772 "num_base_bdevs_operational": 2, 00:09:37.772 "base_bdevs_list": [ 00:09:37.772 { 00:09:37.772 "name": null, 00:09:37.772 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:37.772 "is_configured": false, 00:09:37.772 "data_offset": 0, 00:09:37.772 "data_size": 63488 00:09:37.772 }, 00:09:37.772 { 00:09:37.772 "name": "BaseBdev2", 00:09:37.772 "uuid": "c31a1d78-bcbf-513c-86c0-e30dd5598d17", 00:09:37.772 "is_configured": true, 00:09:37.772 "data_offset": 2048, 00:09:37.772 "data_size": 63488 00:09:37.772 }, 00:09:37.772 { 00:09:37.772 "name": "BaseBdev3", 00:09:37.772 "uuid": "67d69fa3-51d6-5dc6-8ff0-fe245fe0209c", 00:09:37.772 "is_configured": true, 00:09:37.772 "data_offset": 2048, 00:09:37.772 "data_size": 63488 00:09:37.772 } 00:09:37.772 ] 00:09:37.772 }' 00:09:37.772 15:14:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:37.772 15:14:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.342 15:14:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:38.342 15:14:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.342 15:14:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.342 [2024-11-27 15:14:06.216210] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:38.342 [2024-11-27 15:14:06.216297] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:38.342 [2024-11-27 15:14:06.218788] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:38.342 [2024-11-27 15:14:06.218868] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:38.342 [2024-11-27 15:14:06.218980] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:38.342 [2024-11-27 15:14:06.219039] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name raid_bdev1, state offline 00:09:38.342 { 00:09:38.342 "results": [ 00:09:38.342 { 00:09:38.342 "job": "raid_bdev1", 00:09:38.342 "core_mask": "0x1", 00:09:38.342 "workload": "randrw", 00:09:38.342 "percentage": 50, 00:09:38.342 "status": "finished", 00:09:38.342 "queue_depth": 1, 00:09:38.342 "io_size": 131072, 00:09:38.342 "runtime": 1.324117, 00:09:38.342 "iops": 15734.259132689936, 00:09:38.342 "mibps": 1966.782391586242, 00:09:38.342 "io_failed": 0, 00:09:38.342 "io_timeout": 0, 00:09:38.342 "avg_latency_us": 60.94241685052105, 00:09:38.342 "min_latency_us": 23.699563318777294, 00:09:38.342 "max_latency_us": 1359.3711790393013 00:09:38.342 } 00:09:38.342 ], 00:09:38.342 "core_count": 1 00:09:38.342 } 00:09:38.342 15:14:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.342 15:14:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 80391 00:09:38.342 15:14:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 80391 ']' 00:09:38.342 15:14:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 80391 00:09:38.342 15:14:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:09:38.342 15:14:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:38.342 15:14:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 80391 00:09:38.342 15:14:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:38.342 15:14:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:38.342 15:14:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 80391' 00:09:38.342 killing process with pid 80391 00:09:38.342 15:14:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 80391 00:09:38.342 [2024-11-27 15:14:06.268298] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:38.342 15:14:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 80391 00:09:38.342 [2024-11-27 15:14:06.294106] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:38.604 15:14:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.qeNCYOfu8O 00:09:38.604 15:14:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:38.604 15:14:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:38.604 15:14:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:09:38.604 15:14:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:09:38.604 ************************************ 00:09:38.604 END TEST raid_write_error_test 00:09:38.604 ************************************ 00:09:38.604 15:14:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:38.604 15:14:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:38.604 15:14:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:09:38.604 00:09:38.604 real 0m3.172s 00:09:38.604 user 0m3.943s 00:09:38.604 sys 0m0.566s 00:09:38.604 15:14:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:38.604 15:14:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.604 15:14:06 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:09:38.604 15:14:06 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:09:38.604 15:14:06 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 4 false 00:09:38.604 15:14:06 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:38.604 15:14:06 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:38.604 15:14:06 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:38.604 ************************************ 00:09:38.604 START TEST raid_state_function_test 00:09:38.604 ************************************ 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 4 false 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=80518 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 80518' 00:09:38.604 Process raid pid: 80518 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 80518 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 80518 ']' 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:38.604 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:38.604 15:14:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.604 [2024-11-27 15:14:06.688333] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:09:38.604 [2024-11-27 15:14:06.688479] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:38.863 [2024-11-27 15:14:06.853279] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:38.863 [2024-11-27 15:14:06.879412] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:38.863 [2024-11-27 15:14:06.922976] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:38.863 [2024-11-27 15:14:06.923013] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:39.429 15:14:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:39.429 15:14:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:09:39.429 15:14:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:39.429 15:14:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:39.430 15:14:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.430 [2024-11-27 15:14:07.514411] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:39.430 [2024-11-27 15:14:07.514527] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:39.430 [2024-11-27 15:14:07.514559] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:39.430 [2024-11-27 15:14:07.514582] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:39.430 [2024-11-27 15:14:07.514600] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:39.430 [2024-11-27 15:14:07.514622] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:39.430 [2024-11-27 15:14:07.514639] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:39.430 [2024-11-27 15:14:07.514660] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:39.430 15:14:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:39.430 15:14:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:39.430 15:14:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:39.430 15:14:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:39.430 15:14:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:39.430 15:14:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:39.430 15:14:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:39.430 15:14:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:39.430 15:14:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:39.430 15:14:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:39.430 15:14:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:39.430 15:14:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:39.430 15:14:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:39.430 15:14:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:39.430 15:14:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.688 15:14:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:39.688 15:14:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:39.688 "name": "Existed_Raid", 00:09:39.688 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:39.688 "strip_size_kb": 64, 00:09:39.688 "state": "configuring", 00:09:39.688 "raid_level": "raid0", 00:09:39.688 "superblock": false, 00:09:39.688 "num_base_bdevs": 4, 00:09:39.688 "num_base_bdevs_discovered": 0, 00:09:39.688 "num_base_bdevs_operational": 4, 00:09:39.688 "base_bdevs_list": [ 00:09:39.688 { 00:09:39.688 "name": "BaseBdev1", 00:09:39.688 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:39.688 "is_configured": false, 00:09:39.688 "data_offset": 0, 00:09:39.688 "data_size": 0 00:09:39.688 }, 00:09:39.688 { 00:09:39.688 "name": "BaseBdev2", 00:09:39.688 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:39.688 "is_configured": false, 00:09:39.688 "data_offset": 0, 00:09:39.688 "data_size": 0 00:09:39.688 }, 00:09:39.688 { 00:09:39.688 "name": "BaseBdev3", 00:09:39.688 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:39.688 "is_configured": false, 00:09:39.688 "data_offset": 0, 00:09:39.688 "data_size": 0 00:09:39.688 }, 00:09:39.688 { 00:09:39.688 "name": "BaseBdev4", 00:09:39.688 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:39.688 "is_configured": false, 00:09:39.688 "data_offset": 0, 00:09:39.688 "data_size": 0 00:09:39.688 } 00:09:39.688 ] 00:09:39.688 }' 00:09:39.688 15:14:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:39.688 15:14:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.947 15:14:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:39.947 15:14:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:39.947 15:14:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.947 [2024-11-27 15:14:07.929600] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:39.947 [2024-11-27 15:14:07.929683] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:09:39.947 15:14:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:39.947 15:14:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:39.947 15:14:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:39.947 15:14:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.947 [2024-11-27 15:14:07.941598] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:39.947 [2024-11-27 15:14:07.941643] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:39.947 [2024-11-27 15:14:07.941652] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:39.947 [2024-11-27 15:14:07.941662] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:39.947 [2024-11-27 15:14:07.941668] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:39.947 [2024-11-27 15:14:07.941677] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:39.947 [2024-11-27 15:14:07.941683] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:39.947 [2024-11-27 15:14:07.941692] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:39.947 15:14:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:39.947 15:14:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:39.947 15:14:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:39.947 15:14:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.947 [2024-11-27 15:14:07.962663] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:39.947 BaseBdev1 00:09:39.947 15:14:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:39.947 15:14:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:39.947 15:14:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:39.947 15:14:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:39.947 15:14:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:39.947 15:14:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:39.947 15:14:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:39.947 15:14:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:39.947 15:14:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:39.947 15:14:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.947 15:14:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:39.947 15:14:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:39.947 15:14:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:39.947 15:14:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.947 [ 00:09:39.947 { 00:09:39.947 "name": "BaseBdev1", 00:09:39.947 "aliases": [ 00:09:39.947 "3cf2b8a4-ae50-4e32-a441-40ed37a78bc1" 00:09:39.947 ], 00:09:39.947 "product_name": "Malloc disk", 00:09:39.947 "block_size": 512, 00:09:39.947 "num_blocks": 65536, 00:09:39.947 "uuid": "3cf2b8a4-ae50-4e32-a441-40ed37a78bc1", 00:09:39.947 "assigned_rate_limits": { 00:09:39.947 "rw_ios_per_sec": 0, 00:09:39.947 "rw_mbytes_per_sec": 0, 00:09:39.947 "r_mbytes_per_sec": 0, 00:09:39.947 "w_mbytes_per_sec": 0 00:09:39.947 }, 00:09:39.947 "claimed": true, 00:09:39.947 "claim_type": "exclusive_write", 00:09:39.947 "zoned": false, 00:09:39.947 "supported_io_types": { 00:09:39.947 "read": true, 00:09:39.947 "write": true, 00:09:39.947 "unmap": true, 00:09:39.947 "flush": true, 00:09:39.947 "reset": true, 00:09:39.947 "nvme_admin": false, 00:09:39.947 "nvme_io": false, 00:09:39.947 "nvme_io_md": false, 00:09:39.947 "write_zeroes": true, 00:09:39.947 "zcopy": true, 00:09:39.947 "get_zone_info": false, 00:09:39.947 "zone_management": false, 00:09:39.947 "zone_append": false, 00:09:39.947 "compare": false, 00:09:39.947 "compare_and_write": false, 00:09:39.947 "abort": true, 00:09:39.947 "seek_hole": false, 00:09:39.947 "seek_data": false, 00:09:39.947 "copy": true, 00:09:39.947 "nvme_iov_md": false 00:09:39.947 }, 00:09:39.947 "memory_domains": [ 00:09:39.947 { 00:09:39.947 "dma_device_id": "system", 00:09:39.947 "dma_device_type": 1 00:09:39.947 }, 00:09:39.947 { 00:09:39.947 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:39.947 "dma_device_type": 2 00:09:39.947 } 00:09:39.947 ], 00:09:39.947 "driver_specific": {} 00:09:39.947 } 00:09:39.947 ] 00:09:39.947 15:14:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:39.947 15:14:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:39.947 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:39.947 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:39.947 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:39.947 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:39.947 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:39.947 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:39.947 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:39.947 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:39.947 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:39.947 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:39.947 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:39.947 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:39.947 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:39.947 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.947 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.207 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:40.207 "name": "Existed_Raid", 00:09:40.207 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:40.207 "strip_size_kb": 64, 00:09:40.207 "state": "configuring", 00:09:40.207 "raid_level": "raid0", 00:09:40.207 "superblock": false, 00:09:40.207 "num_base_bdevs": 4, 00:09:40.207 "num_base_bdevs_discovered": 1, 00:09:40.207 "num_base_bdevs_operational": 4, 00:09:40.207 "base_bdevs_list": [ 00:09:40.207 { 00:09:40.207 "name": "BaseBdev1", 00:09:40.207 "uuid": "3cf2b8a4-ae50-4e32-a441-40ed37a78bc1", 00:09:40.207 "is_configured": true, 00:09:40.207 "data_offset": 0, 00:09:40.207 "data_size": 65536 00:09:40.207 }, 00:09:40.207 { 00:09:40.207 "name": "BaseBdev2", 00:09:40.207 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:40.207 "is_configured": false, 00:09:40.207 "data_offset": 0, 00:09:40.207 "data_size": 0 00:09:40.207 }, 00:09:40.207 { 00:09:40.207 "name": "BaseBdev3", 00:09:40.207 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:40.207 "is_configured": false, 00:09:40.207 "data_offset": 0, 00:09:40.207 "data_size": 0 00:09:40.207 }, 00:09:40.207 { 00:09:40.207 "name": "BaseBdev4", 00:09:40.207 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:40.207 "is_configured": false, 00:09:40.207 "data_offset": 0, 00:09:40.207 "data_size": 0 00:09:40.207 } 00:09:40.207 ] 00:09:40.207 }' 00:09:40.207 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:40.207 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.467 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:40.467 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.467 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.467 [2024-11-27 15:14:08.385972] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:40.467 [2024-11-27 15:14:08.386072] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:09:40.467 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.467 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:40.467 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.467 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.467 [2024-11-27 15:14:08.397976] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:40.467 [2024-11-27 15:14:08.399835] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:40.467 [2024-11-27 15:14:08.399933] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:40.467 [2024-11-27 15:14:08.399965] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:40.467 [2024-11-27 15:14:08.399990] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:40.467 [2024-11-27 15:14:08.400012] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:40.467 [2024-11-27 15:14:08.400034] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:40.467 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.467 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:40.467 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:40.467 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:40.467 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:40.467 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:40.467 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:40.467 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:40.467 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:40.467 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:40.467 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:40.467 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:40.467 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:40.467 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:40.467 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:40.467 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.467 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.467 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.467 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:40.467 "name": "Existed_Raid", 00:09:40.467 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:40.467 "strip_size_kb": 64, 00:09:40.467 "state": "configuring", 00:09:40.467 "raid_level": "raid0", 00:09:40.467 "superblock": false, 00:09:40.467 "num_base_bdevs": 4, 00:09:40.467 "num_base_bdevs_discovered": 1, 00:09:40.467 "num_base_bdevs_operational": 4, 00:09:40.467 "base_bdevs_list": [ 00:09:40.467 { 00:09:40.467 "name": "BaseBdev1", 00:09:40.467 "uuid": "3cf2b8a4-ae50-4e32-a441-40ed37a78bc1", 00:09:40.467 "is_configured": true, 00:09:40.467 "data_offset": 0, 00:09:40.467 "data_size": 65536 00:09:40.467 }, 00:09:40.467 { 00:09:40.467 "name": "BaseBdev2", 00:09:40.467 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:40.467 "is_configured": false, 00:09:40.467 "data_offset": 0, 00:09:40.467 "data_size": 0 00:09:40.467 }, 00:09:40.467 { 00:09:40.467 "name": "BaseBdev3", 00:09:40.467 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:40.467 "is_configured": false, 00:09:40.467 "data_offset": 0, 00:09:40.468 "data_size": 0 00:09:40.468 }, 00:09:40.468 { 00:09:40.468 "name": "BaseBdev4", 00:09:40.468 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:40.468 "is_configured": false, 00:09:40.468 "data_offset": 0, 00:09:40.468 "data_size": 0 00:09:40.468 } 00:09:40.468 ] 00:09:40.468 }' 00:09:40.468 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:40.468 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.037 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:41.037 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.037 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.037 [2024-11-27 15:14:08.868180] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:41.037 BaseBdev2 00:09:41.037 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.037 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:41.037 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:41.037 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:41.037 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:41.037 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:41.037 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:41.037 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:41.037 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.037 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.037 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.037 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:41.037 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.037 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.037 [ 00:09:41.037 { 00:09:41.037 "name": "BaseBdev2", 00:09:41.037 "aliases": [ 00:09:41.037 "ec03fbaf-e5d0-4254-8589-b91e73c74929" 00:09:41.037 ], 00:09:41.037 "product_name": "Malloc disk", 00:09:41.037 "block_size": 512, 00:09:41.037 "num_blocks": 65536, 00:09:41.037 "uuid": "ec03fbaf-e5d0-4254-8589-b91e73c74929", 00:09:41.037 "assigned_rate_limits": { 00:09:41.037 "rw_ios_per_sec": 0, 00:09:41.037 "rw_mbytes_per_sec": 0, 00:09:41.037 "r_mbytes_per_sec": 0, 00:09:41.037 "w_mbytes_per_sec": 0 00:09:41.037 }, 00:09:41.037 "claimed": true, 00:09:41.037 "claim_type": "exclusive_write", 00:09:41.037 "zoned": false, 00:09:41.037 "supported_io_types": { 00:09:41.037 "read": true, 00:09:41.037 "write": true, 00:09:41.037 "unmap": true, 00:09:41.037 "flush": true, 00:09:41.037 "reset": true, 00:09:41.037 "nvme_admin": false, 00:09:41.037 "nvme_io": false, 00:09:41.037 "nvme_io_md": false, 00:09:41.037 "write_zeroes": true, 00:09:41.037 "zcopy": true, 00:09:41.037 "get_zone_info": false, 00:09:41.037 "zone_management": false, 00:09:41.037 "zone_append": false, 00:09:41.037 "compare": false, 00:09:41.038 "compare_and_write": false, 00:09:41.038 "abort": true, 00:09:41.038 "seek_hole": false, 00:09:41.038 "seek_data": false, 00:09:41.038 "copy": true, 00:09:41.038 "nvme_iov_md": false 00:09:41.038 }, 00:09:41.038 "memory_domains": [ 00:09:41.038 { 00:09:41.038 "dma_device_id": "system", 00:09:41.038 "dma_device_type": 1 00:09:41.038 }, 00:09:41.038 { 00:09:41.038 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:41.038 "dma_device_type": 2 00:09:41.038 } 00:09:41.038 ], 00:09:41.038 "driver_specific": {} 00:09:41.038 } 00:09:41.038 ] 00:09:41.038 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.038 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:41.038 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:41.038 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:41.038 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:41.038 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:41.038 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:41.038 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:41.038 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:41.038 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:41.038 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:41.038 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:41.038 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:41.038 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:41.038 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:41.038 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:41.038 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.038 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.038 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.038 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:41.038 "name": "Existed_Raid", 00:09:41.038 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:41.038 "strip_size_kb": 64, 00:09:41.038 "state": "configuring", 00:09:41.038 "raid_level": "raid0", 00:09:41.038 "superblock": false, 00:09:41.038 "num_base_bdevs": 4, 00:09:41.038 "num_base_bdevs_discovered": 2, 00:09:41.038 "num_base_bdevs_operational": 4, 00:09:41.038 "base_bdevs_list": [ 00:09:41.038 { 00:09:41.038 "name": "BaseBdev1", 00:09:41.038 "uuid": "3cf2b8a4-ae50-4e32-a441-40ed37a78bc1", 00:09:41.038 "is_configured": true, 00:09:41.038 "data_offset": 0, 00:09:41.038 "data_size": 65536 00:09:41.038 }, 00:09:41.038 { 00:09:41.038 "name": "BaseBdev2", 00:09:41.038 "uuid": "ec03fbaf-e5d0-4254-8589-b91e73c74929", 00:09:41.038 "is_configured": true, 00:09:41.038 "data_offset": 0, 00:09:41.038 "data_size": 65536 00:09:41.038 }, 00:09:41.038 { 00:09:41.038 "name": "BaseBdev3", 00:09:41.038 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:41.038 "is_configured": false, 00:09:41.038 "data_offset": 0, 00:09:41.038 "data_size": 0 00:09:41.038 }, 00:09:41.038 { 00:09:41.038 "name": "BaseBdev4", 00:09:41.038 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:41.038 "is_configured": false, 00:09:41.038 "data_offset": 0, 00:09:41.038 "data_size": 0 00:09:41.038 } 00:09:41.038 ] 00:09:41.038 }' 00:09:41.038 15:14:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:41.038 15:14:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.297 [2024-11-27 15:14:09.340124] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:41.297 BaseBdev3 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.297 [ 00:09:41.297 { 00:09:41.297 "name": "BaseBdev3", 00:09:41.297 "aliases": [ 00:09:41.297 "a0e20c97-a52b-4f32-a85c-116bcc87873f" 00:09:41.297 ], 00:09:41.297 "product_name": "Malloc disk", 00:09:41.297 "block_size": 512, 00:09:41.297 "num_blocks": 65536, 00:09:41.297 "uuid": "a0e20c97-a52b-4f32-a85c-116bcc87873f", 00:09:41.297 "assigned_rate_limits": { 00:09:41.297 "rw_ios_per_sec": 0, 00:09:41.297 "rw_mbytes_per_sec": 0, 00:09:41.297 "r_mbytes_per_sec": 0, 00:09:41.297 "w_mbytes_per_sec": 0 00:09:41.297 }, 00:09:41.297 "claimed": true, 00:09:41.297 "claim_type": "exclusive_write", 00:09:41.297 "zoned": false, 00:09:41.297 "supported_io_types": { 00:09:41.297 "read": true, 00:09:41.297 "write": true, 00:09:41.297 "unmap": true, 00:09:41.297 "flush": true, 00:09:41.297 "reset": true, 00:09:41.297 "nvme_admin": false, 00:09:41.297 "nvme_io": false, 00:09:41.297 "nvme_io_md": false, 00:09:41.297 "write_zeroes": true, 00:09:41.297 "zcopy": true, 00:09:41.297 "get_zone_info": false, 00:09:41.297 "zone_management": false, 00:09:41.297 "zone_append": false, 00:09:41.297 "compare": false, 00:09:41.297 "compare_and_write": false, 00:09:41.297 "abort": true, 00:09:41.297 "seek_hole": false, 00:09:41.297 "seek_data": false, 00:09:41.297 "copy": true, 00:09:41.297 "nvme_iov_md": false 00:09:41.297 }, 00:09:41.297 "memory_domains": [ 00:09:41.297 { 00:09:41.297 "dma_device_id": "system", 00:09:41.297 "dma_device_type": 1 00:09:41.297 }, 00:09:41.297 { 00:09:41.297 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:41.297 "dma_device_type": 2 00:09:41.297 } 00:09:41.297 ], 00:09:41.297 "driver_specific": {} 00:09:41.297 } 00:09:41.297 ] 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.297 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.556 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.556 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:41.556 "name": "Existed_Raid", 00:09:41.556 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:41.556 "strip_size_kb": 64, 00:09:41.556 "state": "configuring", 00:09:41.556 "raid_level": "raid0", 00:09:41.556 "superblock": false, 00:09:41.556 "num_base_bdevs": 4, 00:09:41.556 "num_base_bdevs_discovered": 3, 00:09:41.556 "num_base_bdevs_operational": 4, 00:09:41.556 "base_bdevs_list": [ 00:09:41.556 { 00:09:41.556 "name": "BaseBdev1", 00:09:41.556 "uuid": "3cf2b8a4-ae50-4e32-a441-40ed37a78bc1", 00:09:41.556 "is_configured": true, 00:09:41.556 "data_offset": 0, 00:09:41.556 "data_size": 65536 00:09:41.556 }, 00:09:41.556 { 00:09:41.556 "name": "BaseBdev2", 00:09:41.556 "uuid": "ec03fbaf-e5d0-4254-8589-b91e73c74929", 00:09:41.556 "is_configured": true, 00:09:41.556 "data_offset": 0, 00:09:41.556 "data_size": 65536 00:09:41.556 }, 00:09:41.556 { 00:09:41.556 "name": "BaseBdev3", 00:09:41.556 "uuid": "a0e20c97-a52b-4f32-a85c-116bcc87873f", 00:09:41.556 "is_configured": true, 00:09:41.556 "data_offset": 0, 00:09:41.556 "data_size": 65536 00:09:41.556 }, 00:09:41.556 { 00:09:41.556 "name": "BaseBdev4", 00:09:41.556 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:41.556 "is_configured": false, 00:09:41.556 "data_offset": 0, 00:09:41.556 "data_size": 0 00:09:41.556 } 00:09:41.556 ] 00:09:41.556 }' 00:09:41.556 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:41.556 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.815 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:41.815 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.815 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.815 [2024-11-27 15:14:09.874408] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:41.815 [2024-11-27 15:14:09.874530] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:09:41.815 [2024-11-27 15:14:09.874556] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:09:41.815 [2024-11-27 15:14:09.874838] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:09:41.815 [2024-11-27 15:14:09.875032] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:09:41.815 [2024-11-27 15:14:09.875080] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:09:41.815 [2024-11-27 15:14:09.875311] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:41.815 BaseBdev4 00:09:41.815 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.815 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:09:41.815 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:09:41.815 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:41.815 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:41.815 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:41.815 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:41.815 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:41.815 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.815 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.815 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.815 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:41.815 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.815 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.815 [ 00:09:41.815 { 00:09:41.815 "name": "BaseBdev4", 00:09:41.815 "aliases": [ 00:09:41.815 "f9306a1c-56a6-44a8-8294-ad6ae21ac6ab" 00:09:41.815 ], 00:09:41.815 "product_name": "Malloc disk", 00:09:41.815 "block_size": 512, 00:09:41.815 "num_blocks": 65536, 00:09:41.815 "uuid": "f9306a1c-56a6-44a8-8294-ad6ae21ac6ab", 00:09:41.815 "assigned_rate_limits": { 00:09:41.815 "rw_ios_per_sec": 0, 00:09:41.815 "rw_mbytes_per_sec": 0, 00:09:41.815 "r_mbytes_per_sec": 0, 00:09:41.815 "w_mbytes_per_sec": 0 00:09:41.815 }, 00:09:41.815 "claimed": true, 00:09:41.815 "claim_type": "exclusive_write", 00:09:41.815 "zoned": false, 00:09:41.815 "supported_io_types": { 00:09:41.815 "read": true, 00:09:41.815 "write": true, 00:09:41.815 "unmap": true, 00:09:41.815 "flush": true, 00:09:41.815 "reset": true, 00:09:41.815 "nvme_admin": false, 00:09:41.815 "nvme_io": false, 00:09:41.815 "nvme_io_md": false, 00:09:41.815 "write_zeroes": true, 00:09:41.815 "zcopy": true, 00:09:41.815 "get_zone_info": false, 00:09:41.815 "zone_management": false, 00:09:41.815 "zone_append": false, 00:09:41.815 "compare": false, 00:09:41.815 "compare_and_write": false, 00:09:41.815 "abort": true, 00:09:41.815 "seek_hole": false, 00:09:41.815 "seek_data": false, 00:09:41.815 "copy": true, 00:09:41.815 "nvme_iov_md": false 00:09:41.815 }, 00:09:41.815 "memory_domains": [ 00:09:41.815 { 00:09:41.815 "dma_device_id": "system", 00:09:41.815 "dma_device_type": 1 00:09:41.815 }, 00:09:41.815 { 00:09:41.815 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:41.815 "dma_device_type": 2 00:09:41.815 } 00:09:41.815 ], 00:09:41.816 "driver_specific": {} 00:09:41.816 } 00:09:41.816 ] 00:09:41.816 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.816 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:41.816 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:41.816 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:41.816 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:09:41.816 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:41.816 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:41.816 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:41.816 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:41.816 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:41.816 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:41.816 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:41.816 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:41.816 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:41.816 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:41.816 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.816 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.074 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:42.074 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:42.074 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:42.074 "name": "Existed_Raid", 00:09:42.074 "uuid": "61a067ad-6d49-4113-904e-37fb3998e6d6", 00:09:42.074 "strip_size_kb": 64, 00:09:42.074 "state": "online", 00:09:42.074 "raid_level": "raid0", 00:09:42.074 "superblock": false, 00:09:42.074 "num_base_bdevs": 4, 00:09:42.074 "num_base_bdevs_discovered": 4, 00:09:42.074 "num_base_bdevs_operational": 4, 00:09:42.074 "base_bdevs_list": [ 00:09:42.074 { 00:09:42.074 "name": "BaseBdev1", 00:09:42.074 "uuid": "3cf2b8a4-ae50-4e32-a441-40ed37a78bc1", 00:09:42.074 "is_configured": true, 00:09:42.074 "data_offset": 0, 00:09:42.074 "data_size": 65536 00:09:42.074 }, 00:09:42.074 { 00:09:42.074 "name": "BaseBdev2", 00:09:42.074 "uuid": "ec03fbaf-e5d0-4254-8589-b91e73c74929", 00:09:42.074 "is_configured": true, 00:09:42.074 "data_offset": 0, 00:09:42.074 "data_size": 65536 00:09:42.074 }, 00:09:42.074 { 00:09:42.074 "name": "BaseBdev3", 00:09:42.074 "uuid": "a0e20c97-a52b-4f32-a85c-116bcc87873f", 00:09:42.074 "is_configured": true, 00:09:42.074 "data_offset": 0, 00:09:42.074 "data_size": 65536 00:09:42.074 }, 00:09:42.074 { 00:09:42.074 "name": "BaseBdev4", 00:09:42.074 "uuid": "f9306a1c-56a6-44a8-8294-ad6ae21ac6ab", 00:09:42.074 "is_configured": true, 00:09:42.074 "data_offset": 0, 00:09:42.074 "data_size": 65536 00:09:42.074 } 00:09:42.074 ] 00:09:42.074 }' 00:09:42.074 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:42.074 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.332 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:42.332 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:42.332 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:42.332 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:42.332 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:42.332 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:42.332 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:42.332 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:42.332 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:42.332 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.332 [2024-11-27 15:14:10.334026] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:42.332 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:42.332 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:42.332 "name": "Existed_Raid", 00:09:42.332 "aliases": [ 00:09:42.332 "61a067ad-6d49-4113-904e-37fb3998e6d6" 00:09:42.332 ], 00:09:42.332 "product_name": "Raid Volume", 00:09:42.332 "block_size": 512, 00:09:42.332 "num_blocks": 262144, 00:09:42.332 "uuid": "61a067ad-6d49-4113-904e-37fb3998e6d6", 00:09:42.332 "assigned_rate_limits": { 00:09:42.332 "rw_ios_per_sec": 0, 00:09:42.332 "rw_mbytes_per_sec": 0, 00:09:42.332 "r_mbytes_per_sec": 0, 00:09:42.332 "w_mbytes_per_sec": 0 00:09:42.332 }, 00:09:42.332 "claimed": false, 00:09:42.332 "zoned": false, 00:09:42.332 "supported_io_types": { 00:09:42.332 "read": true, 00:09:42.332 "write": true, 00:09:42.332 "unmap": true, 00:09:42.332 "flush": true, 00:09:42.332 "reset": true, 00:09:42.332 "nvme_admin": false, 00:09:42.332 "nvme_io": false, 00:09:42.332 "nvme_io_md": false, 00:09:42.332 "write_zeroes": true, 00:09:42.332 "zcopy": false, 00:09:42.332 "get_zone_info": false, 00:09:42.332 "zone_management": false, 00:09:42.332 "zone_append": false, 00:09:42.332 "compare": false, 00:09:42.332 "compare_and_write": false, 00:09:42.332 "abort": false, 00:09:42.332 "seek_hole": false, 00:09:42.332 "seek_data": false, 00:09:42.332 "copy": false, 00:09:42.332 "nvme_iov_md": false 00:09:42.332 }, 00:09:42.332 "memory_domains": [ 00:09:42.332 { 00:09:42.332 "dma_device_id": "system", 00:09:42.332 "dma_device_type": 1 00:09:42.332 }, 00:09:42.332 { 00:09:42.332 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:42.332 "dma_device_type": 2 00:09:42.332 }, 00:09:42.332 { 00:09:42.332 "dma_device_id": "system", 00:09:42.332 "dma_device_type": 1 00:09:42.332 }, 00:09:42.332 { 00:09:42.332 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:42.332 "dma_device_type": 2 00:09:42.332 }, 00:09:42.332 { 00:09:42.332 "dma_device_id": "system", 00:09:42.332 "dma_device_type": 1 00:09:42.332 }, 00:09:42.332 { 00:09:42.332 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:42.332 "dma_device_type": 2 00:09:42.332 }, 00:09:42.332 { 00:09:42.332 "dma_device_id": "system", 00:09:42.332 "dma_device_type": 1 00:09:42.332 }, 00:09:42.332 { 00:09:42.332 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:42.332 "dma_device_type": 2 00:09:42.332 } 00:09:42.332 ], 00:09:42.332 "driver_specific": { 00:09:42.332 "raid": { 00:09:42.332 "uuid": "61a067ad-6d49-4113-904e-37fb3998e6d6", 00:09:42.332 "strip_size_kb": 64, 00:09:42.332 "state": "online", 00:09:42.332 "raid_level": "raid0", 00:09:42.332 "superblock": false, 00:09:42.332 "num_base_bdevs": 4, 00:09:42.332 "num_base_bdevs_discovered": 4, 00:09:42.332 "num_base_bdevs_operational": 4, 00:09:42.332 "base_bdevs_list": [ 00:09:42.332 { 00:09:42.332 "name": "BaseBdev1", 00:09:42.332 "uuid": "3cf2b8a4-ae50-4e32-a441-40ed37a78bc1", 00:09:42.332 "is_configured": true, 00:09:42.332 "data_offset": 0, 00:09:42.332 "data_size": 65536 00:09:42.332 }, 00:09:42.332 { 00:09:42.332 "name": "BaseBdev2", 00:09:42.332 "uuid": "ec03fbaf-e5d0-4254-8589-b91e73c74929", 00:09:42.332 "is_configured": true, 00:09:42.332 "data_offset": 0, 00:09:42.332 "data_size": 65536 00:09:42.332 }, 00:09:42.332 { 00:09:42.332 "name": "BaseBdev3", 00:09:42.332 "uuid": "a0e20c97-a52b-4f32-a85c-116bcc87873f", 00:09:42.332 "is_configured": true, 00:09:42.332 "data_offset": 0, 00:09:42.333 "data_size": 65536 00:09:42.333 }, 00:09:42.333 { 00:09:42.333 "name": "BaseBdev4", 00:09:42.333 "uuid": "f9306a1c-56a6-44a8-8294-ad6ae21ac6ab", 00:09:42.333 "is_configured": true, 00:09:42.333 "data_offset": 0, 00:09:42.333 "data_size": 65536 00:09:42.333 } 00:09:42.333 ] 00:09:42.333 } 00:09:42.333 } 00:09:42.333 }' 00:09:42.333 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:42.333 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:42.333 BaseBdev2 00:09:42.333 BaseBdev3 00:09:42.333 BaseBdev4' 00:09:42.333 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:42.333 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:42.333 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:42.333 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:42.333 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:42.333 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.333 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:42.592 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:42.592 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.593 [2024-11-27 15:14:10.621242] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:42.593 [2024-11-27 15:14:10.621317] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:42.593 [2024-11-27 15:14:10.621392] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 3 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:42.593 "name": "Existed_Raid", 00:09:42.593 "uuid": "61a067ad-6d49-4113-904e-37fb3998e6d6", 00:09:42.593 "strip_size_kb": 64, 00:09:42.593 "state": "offline", 00:09:42.593 "raid_level": "raid0", 00:09:42.593 "superblock": false, 00:09:42.593 "num_base_bdevs": 4, 00:09:42.593 "num_base_bdevs_discovered": 3, 00:09:42.593 "num_base_bdevs_operational": 3, 00:09:42.593 "base_bdevs_list": [ 00:09:42.593 { 00:09:42.593 "name": null, 00:09:42.593 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:42.593 "is_configured": false, 00:09:42.593 "data_offset": 0, 00:09:42.593 "data_size": 65536 00:09:42.593 }, 00:09:42.593 { 00:09:42.593 "name": "BaseBdev2", 00:09:42.593 "uuid": "ec03fbaf-e5d0-4254-8589-b91e73c74929", 00:09:42.593 "is_configured": true, 00:09:42.593 "data_offset": 0, 00:09:42.593 "data_size": 65536 00:09:42.593 }, 00:09:42.593 { 00:09:42.593 "name": "BaseBdev3", 00:09:42.593 "uuid": "a0e20c97-a52b-4f32-a85c-116bcc87873f", 00:09:42.593 "is_configured": true, 00:09:42.593 "data_offset": 0, 00:09:42.593 "data_size": 65536 00:09:42.593 }, 00:09:42.593 { 00:09:42.593 "name": "BaseBdev4", 00:09:42.593 "uuid": "f9306a1c-56a6-44a8-8294-ad6ae21ac6ab", 00:09:42.593 "is_configured": true, 00:09:42.593 "data_offset": 0, 00:09:42.593 "data_size": 65536 00:09:42.593 } 00:09:42.593 ] 00:09:42.593 }' 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:42.593 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.163 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:43.163 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:43.163 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:43.163 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:43.163 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.163 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.163 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.163 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:43.163 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:43.163 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:43.163 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.163 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.163 [2024-11-27 15:14:11.163833] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:43.163 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.163 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:43.163 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:43.163 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:43.163 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:43.163 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.163 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.163 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.163 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:43.163 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:43.163 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:43.163 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.163 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.163 [2024-11-27 15:14:11.230998] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:43.163 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.163 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:43.163 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:43.163 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:43.163 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:43.163 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.163 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.424 [2024-11-27 15:14:11.298029] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:09:43.424 [2024-11-27 15:14:11.298118] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.424 BaseBdev2 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.424 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.424 [ 00:09:43.424 { 00:09:43.424 "name": "BaseBdev2", 00:09:43.424 "aliases": [ 00:09:43.424 "56be89cb-98c9-4fed-b3b8-32dc061d606a" 00:09:43.424 ], 00:09:43.424 "product_name": "Malloc disk", 00:09:43.424 "block_size": 512, 00:09:43.424 "num_blocks": 65536, 00:09:43.424 "uuid": "56be89cb-98c9-4fed-b3b8-32dc061d606a", 00:09:43.424 "assigned_rate_limits": { 00:09:43.424 "rw_ios_per_sec": 0, 00:09:43.424 "rw_mbytes_per_sec": 0, 00:09:43.425 "r_mbytes_per_sec": 0, 00:09:43.425 "w_mbytes_per_sec": 0 00:09:43.425 }, 00:09:43.425 "claimed": false, 00:09:43.425 "zoned": false, 00:09:43.425 "supported_io_types": { 00:09:43.425 "read": true, 00:09:43.425 "write": true, 00:09:43.425 "unmap": true, 00:09:43.425 "flush": true, 00:09:43.425 "reset": true, 00:09:43.425 "nvme_admin": false, 00:09:43.425 "nvme_io": false, 00:09:43.425 "nvme_io_md": false, 00:09:43.425 "write_zeroes": true, 00:09:43.425 "zcopy": true, 00:09:43.425 "get_zone_info": false, 00:09:43.425 "zone_management": false, 00:09:43.425 "zone_append": false, 00:09:43.425 "compare": false, 00:09:43.425 "compare_and_write": false, 00:09:43.425 "abort": true, 00:09:43.425 "seek_hole": false, 00:09:43.425 "seek_data": false, 00:09:43.425 "copy": true, 00:09:43.425 "nvme_iov_md": false 00:09:43.425 }, 00:09:43.425 "memory_domains": [ 00:09:43.425 { 00:09:43.425 "dma_device_id": "system", 00:09:43.425 "dma_device_type": 1 00:09:43.425 }, 00:09:43.425 { 00:09:43.425 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:43.425 "dma_device_type": 2 00:09:43.425 } 00:09:43.425 ], 00:09:43.425 "driver_specific": {} 00:09:43.425 } 00:09:43.425 ] 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.425 BaseBdev3 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.425 [ 00:09:43.425 { 00:09:43.425 "name": "BaseBdev3", 00:09:43.425 "aliases": [ 00:09:43.425 "22b62ba9-36ce-42d7-94d7-85cf99756e33" 00:09:43.425 ], 00:09:43.425 "product_name": "Malloc disk", 00:09:43.425 "block_size": 512, 00:09:43.425 "num_blocks": 65536, 00:09:43.425 "uuid": "22b62ba9-36ce-42d7-94d7-85cf99756e33", 00:09:43.425 "assigned_rate_limits": { 00:09:43.425 "rw_ios_per_sec": 0, 00:09:43.425 "rw_mbytes_per_sec": 0, 00:09:43.425 "r_mbytes_per_sec": 0, 00:09:43.425 "w_mbytes_per_sec": 0 00:09:43.425 }, 00:09:43.425 "claimed": false, 00:09:43.425 "zoned": false, 00:09:43.425 "supported_io_types": { 00:09:43.425 "read": true, 00:09:43.425 "write": true, 00:09:43.425 "unmap": true, 00:09:43.425 "flush": true, 00:09:43.425 "reset": true, 00:09:43.425 "nvme_admin": false, 00:09:43.425 "nvme_io": false, 00:09:43.425 "nvme_io_md": false, 00:09:43.425 "write_zeroes": true, 00:09:43.425 "zcopy": true, 00:09:43.425 "get_zone_info": false, 00:09:43.425 "zone_management": false, 00:09:43.425 "zone_append": false, 00:09:43.425 "compare": false, 00:09:43.425 "compare_and_write": false, 00:09:43.425 "abort": true, 00:09:43.425 "seek_hole": false, 00:09:43.425 "seek_data": false, 00:09:43.425 "copy": true, 00:09:43.425 "nvme_iov_md": false 00:09:43.425 }, 00:09:43.425 "memory_domains": [ 00:09:43.425 { 00:09:43.425 "dma_device_id": "system", 00:09:43.425 "dma_device_type": 1 00:09:43.425 }, 00:09:43.425 { 00:09:43.425 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:43.425 "dma_device_type": 2 00:09:43.425 } 00:09:43.425 ], 00:09:43.425 "driver_specific": {} 00:09:43.425 } 00:09:43.425 ] 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.425 BaseBdev4 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.425 [ 00:09:43.425 { 00:09:43.425 "name": "BaseBdev4", 00:09:43.425 "aliases": [ 00:09:43.425 "fb647cf8-0896-4972-b9c9-806e17972047" 00:09:43.425 ], 00:09:43.425 "product_name": "Malloc disk", 00:09:43.425 "block_size": 512, 00:09:43.425 "num_blocks": 65536, 00:09:43.425 "uuid": "fb647cf8-0896-4972-b9c9-806e17972047", 00:09:43.425 "assigned_rate_limits": { 00:09:43.425 "rw_ios_per_sec": 0, 00:09:43.425 "rw_mbytes_per_sec": 0, 00:09:43.425 "r_mbytes_per_sec": 0, 00:09:43.425 "w_mbytes_per_sec": 0 00:09:43.425 }, 00:09:43.425 "claimed": false, 00:09:43.425 "zoned": false, 00:09:43.425 "supported_io_types": { 00:09:43.425 "read": true, 00:09:43.425 "write": true, 00:09:43.425 "unmap": true, 00:09:43.425 "flush": true, 00:09:43.425 "reset": true, 00:09:43.425 "nvme_admin": false, 00:09:43.425 "nvme_io": false, 00:09:43.425 "nvme_io_md": false, 00:09:43.425 "write_zeroes": true, 00:09:43.425 "zcopy": true, 00:09:43.425 "get_zone_info": false, 00:09:43.425 "zone_management": false, 00:09:43.425 "zone_append": false, 00:09:43.425 "compare": false, 00:09:43.425 "compare_and_write": false, 00:09:43.425 "abort": true, 00:09:43.425 "seek_hole": false, 00:09:43.425 "seek_data": false, 00:09:43.425 "copy": true, 00:09:43.425 "nvme_iov_md": false 00:09:43.425 }, 00:09:43.425 "memory_domains": [ 00:09:43.425 { 00:09:43.425 "dma_device_id": "system", 00:09:43.425 "dma_device_type": 1 00:09:43.425 }, 00:09:43.425 { 00:09:43.425 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:43.425 "dma_device_type": 2 00:09:43.425 } 00:09:43.425 ], 00:09:43.425 "driver_specific": {} 00:09:43.425 } 00:09:43.425 ] 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.425 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.425 [2024-11-27 15:14:11.507050] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:43.425 [2024-11-27 15:14:11.507169] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:43.425 [2024-11-27 15:14:11.507213] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:43.425 [2024-11-27 15:14:11.509092] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:43.425 [2024-11-27 15:14:11.509190] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:43.426 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.426 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:43.426 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:43.426 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:43.426 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:43.426 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:43.426 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:43.426 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:43.426 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:43.426 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:43.426 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:43.426 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:43.426 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:43.426 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.426 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.685 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.685 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:43.685 "name": "Existed_Raid", 00:09:43.685 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:43.685 "strip_size_kb": 64, 00:09:43.685 "state": "configuring", 00:09:43.685 "raid_level": "raid0", 00:09:43.685 "superblock": false, 00:09:43.685 "num_base_bdevs": 4, 00:09:43.685 "num_base_bdevs_discovered": 3, 00:09:43.685 "num_base_bdevs_operational": 4, 00:09:43.685 "base_bdevs_list": [ 00:09:43.685 { 00:09:43.685 "name": "BaseBdev1", 00:09:43.685 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:43.685 "is_configured": false, 00:09:43.685 "data_offset": 0, 00:09:43.685 "data_size": 0 00:09:43.685 }, 00:09:43.685 { 00:09:43.685 "name": "BaseBdev2", 00:09:43.685 "uuid": "56be89cb-98c9-4fed-b3b8-32dc061d606a", 00:09:43.685 "is_configured": true, 00:09:43.685 "data_offset": 0, 00:09:43.685 "data_size": 65536 00:09:43.685 }, 00:09:43.685 { 00:09:43.685 "name": "BaseBdev3", 00:09:43.685 "uuid": "22b62ba9-36ce-42d7-94d7-85cf99756e33", 00:09:43.685 "is_configured": true, 00:09:43.685 "data_offset": 0, 00:09:43.685 "data_size": 65536 00:09:43.685 }, 00:09:43.685 { 00:09:43.685 "name": "BaseBdev4", 00:09:43.685 "uuid": "fb647cf8-0896-4972-b9c9-806e17972047", 00:09:43.685 "is_configured": true, 00:09:43.685 "data_offset": 0, 00:09:43.685 "data_size": 65536 00:09:43.685 } 00:09:43.685 ] 00:09:43.685 }' 00:09:43.685 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:43.685 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.945 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:43.945 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.945 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.945 [2024-11-27 15:14:11.914382] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:43.945 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.945 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:43.945 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:43.945 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:43.945 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:43.945 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:43.945 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:43.945 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:43.945 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:43.945 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:43.945 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:43.945 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:43.945 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:43.945 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.945 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.945 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.945 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:43.945 "name": "Existed_Raid", 00:09:43.945 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:43.945 "strip_size_kb": 64, 00:09:43.945 "state": "configuring", 00:09:43.945 "raid_level": "raid0", 00:09:43.945 "superblock": false, 00:09:43.945 "num_base_bdevs": 4, 00:09:43.945 "num_base_bdevs_discovered": 2, 00:09:43.945 "num_base_bdevs_operational": 4, 00:09:43.945 "base_bdevs_list": [ 00:09:43.945 { 00:09:43.945 "name": "BaseBdev1", 00:09:43.945 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:43.945 "is_configured": false, 00:09:43.945 "data_offset": 0, 00:09:43.945 "data_size": 0 00:09:43.945 }, 00:09:43.945 { 00:09:43.945 "name": null, 00:09:43.945 "uuid": "56be89cb-98c9-4fed-b3b8-32dc061d606a", 00:09:43.945 "is_configured": false, 00:09:43.945 "data_offset": 0, 00:09:43.945 "data_size": 65536 00:09:43.945 }, 00:09:43.945 { 00:09:43.945 "name": "BaseBdev3", 00:09:43.945 "uuid": "22b62ba9-36ce-42d7-94d7-85cf99756e33", 00:09:43.945 "is_configured": true, 00:09:43.945 "data_offset": 0, 00:09:43.945 "data_size": 65536 00:09:43.945 }, 00:09:43.945 { 00:09:43.945 "name": "BaseBdev4", 00:09:43.945 "uuid": "fb647cf8-0896-4972-b9c9-806e17972047", 00:09:43.945 "is_configured": true, 00:09:43.945 "data_offset": 0, 00:09:43.945 "data_size": 65536 00:09:43.945 } 00:09:43.945 ] 00:09:43.945 }' 00:09:43.945 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:43.945 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.204 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:44.204 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.204 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.204 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.465 BaseBdev1 00:09:44.465 [2024-11-27 15:14:12.360556] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.465 [ 00:09:44.465 { 00:09:44.465 "name": "BaseBdev1", 00:09:44.465 "aliases": [ 00:09:44.465 "471a3f15-cfbf-463d-9206-ae2497843ab5" 00:09:44.465 ], 00:09:44.465 "product_name": "Malloc disk", 00:09:44.465 "block_size": 512, 00:09:44.465 "num_blocks": 65536, 00:09:44.465 "uuid": "471a3f15-cfbf-463d-9206-ae2497843ab5", 00:09:44.465 "assigned_rate_limits": { 00:09:44.465 "rw_ios_per_sec": 0, 00:09:44.465 "rw_mbytes_per_sec": 0, 00:09:44.465 "r_mbytes_per_sec": 0, 00:09:44.465 "w_mbytes_per_sec": 0 00:09:44.465 }, 00:09:44.465 "claimed": true, 00:09:44.465 "claim_type": "exclusive_write", 00:09:44.465 "zoned": false, 00:09:44.465 "supported_io_types": { 00:09:44.465 "read": true, 00:09:44.465 "write": true, 00:09:44.465 "unmap": true, 00:09:44.465 "flush": true, 00:09:44.465 "reset": true, 00:09:44.465 "nvme_admin": false, 00:09:44.465 "nvme_io": false, 00:09:44.465 "nvme_io_md": false, 00:09:44.465 "write_zeroes": true, 00:09:44.465 "zcopy": true, 00:09:44.465 "get_zone_info": false, 00:09:44.465 "zone_management": false, 00:09:44.465 "zone_append": false, 00:09:44.465 "compare": false, 00:09:44.465 "compare_and_write": false, 00:09:44.465 "abort": true, 00:09:44.465 "seek_hole": false, 00:09:44.465 "seek_data": false, 00:09:44.465 "copy": true, 00:09:44.465 "nvme_iov_md": false 00:09:44.465 }, 00:09:44.465 "memory_domains": [ 00:09:44.465 { 00:09:44.465 "dma_device_id": "system", 00:09:44.465 "dma_device_type": 1 00:09:44.465 }, 00:09:44.465 { 00:09:44.465 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:44.465 "dma_device_type": 2 00:09:44.465 } 00:09:44.465 ], 00:09:44.465 "driver_specific": {} 00:09:44.465 } 00:09:44.465 ] 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:44.465 "name": "Existed_Raid", 00:09:44.465 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:44.465 "strip_size_kb": 64, 00:09:44.465 "state": "configuring", 00:09:44.465 "raid_level": "raid0", 00:09:44.465 "superblock": false, 00:09:44.465 "num_base_bdevs": 4, 00:09:44.465 "num_base_bdevs_discovered": 3, 00:09:44.465 "num_base_bdevs_operational": 4, 00:09:44.465 "base_bdevs_list": [ 00:09:44.465 { 00:09:44.465 "name": "BaseBdev1", 00:09:44.465 "uuid": "471a3f15-cfbf-463d-9206-ae2497843ab5", 00:09:44.465 "is_configured": true, 00:09:44.465 "data_offset": 0, 00:09:44.465 "data_size": 65536 00:09:44.465 }, 00:09:44.465 { 00:09:44.465 "name": null, 00:09:44.465 "uuid": "56be89cb-98c9-4fed-b3b8-32dc061d606a", 00:09:44.465 "is_configured": false, 00:09:44.465 "data_offset": 0, 00:09:44.465 "data_size": 65536 00:09:44.465 }, 00:09:44.465 { 00:09:44.465 "name": "BaseBdev3", 00:09:44.465 "uuid": "22b62ba9-36ce-42d7-94d7-85cf99756e33", 00:09:44.465 "is_configured": true, 00:09:44.465 "data_offset": 0, 00:09:44.465 "data_size": 65536 00:09:44.465 }, 00:09:44.465 { 00:09:44.465 "name": "BaseBdev4", 00:09:44.465 "uuid": "fb647cf8-0896-4972-b9c9-806e17972047", 00:09:44.465 "is_configured": true, 00:09:44.465 "data_offset": 0, 00:09:44.465 "data_size": 65536 00:09:44.465 } 00:09:44.465 ] 00:09:44.465 }' 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:44.465 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.725 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:44.725 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:44.725 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.725 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.725 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.985 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:44.985 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:44.985 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.985 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.985 [2024-11-27 15:14:12.847787] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:44.985 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.985 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:44.985 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:44.985 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:44.985 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:44.985 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:44.985 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:44.985 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:44.985 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:44.985 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:44.985 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:44.985 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:44.985 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:44.985 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.985 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.985 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.985 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:44.985 "name": "Existed_Raid", 00:09:44.985 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:44.985 "strip_size_kb": 64, 00:09:44.985 "state": "configuring", 00:09:44.985 "raid_level": "raid0", 00:09:44.985 "superblock": false, 00:09:44.985 "num_base_bdevs": 4, 00:09:44.985 "num_base_bdevs_discovered": 2, 00:09:44.985 "num_base_bdevs_operational": 4, 00:09:44.985 "base_bdevs_list": [ 00:09:44.985 { 00:09:44.985 "name": "BaseBdev1", 00:09:44.985 "uuid": "471a3f15-cfbf-463d-9206-ae2497843ab5", 00:09:44.985 "is_configured": true, 00:09:44.985 "data_offset": 0, 00:09:44.985 "data_size": 65536 00:09:44.985 }, 00:09:44.985 { 00:09:44.985 "name": null, 00:09:44.985 "uuid": "56be89cb-98c9-4fed-b3b8-32dc061d606a", 00:09:44.985 "is_configured": false, 00:09:44.985 "data_offset": 0, 00:09:44.985 "data_size": 65536 00:09:44.985 }, 00:09:44.985 { 00:09:44.985 "name": null, 00:09:44.985 "uuid": "22b62ba9-36ce-42d7-94d7-85cf99756e33", 00:09:44.985 "is_configured": false, 00:09:44.985 "data_offset": 0, 00:09:44.985 "data_size": 65536 00:09:44.985 }, 00:09:44.985 { 00:09:44.985 "name": "BaseBdev4", 00:09:44.985 "uuid": "fb647cf8-0896-4972-b9c9-806e17972047", 00:09:44.985 "is_configured": true, 00:09:44.985 "data_offset": 0, 00:09:44.985 "data_size": 65536 00:09:44.985 } 00:09:44.985 ] 00:09:44.985 }' 00:09:44.985 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:44.985 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.245 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:45.245 15:14:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.245 15:14:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.245 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:45.245 15:14:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.245 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:45.245 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:45.245 15:14:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.245 15:14:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.245 [2024-11-27 15:14:13.323042] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:45.245 15:14:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.245 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:45.245 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:45.245 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:45.245 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:45.245 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:45.245 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:45.245 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:45.245 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:45.245 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:45.245 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:45.245 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:45.245 15:14:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.245 15:14:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.246 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:45.246 15:14:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.515 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:45.515 "name": "Existed_Raid", 00:09:45.515 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:45.515 "strip_size_kb": 64, 00:09:45.515 "state": "configuring", 00:09:45.515 "raid_level": "raid0", 00:09:45.515 "superblock": false, 00:09:45.515 "num_base_bdevs": 4, 00:09:45.515 "num_base_bdevs_discovered": 3, 00:09:45.515 "num_base_bdevs_operational": 4, 00:09:45.515 "base_bdevs_list": [ 00:09:45.515 { 00:09:45.515 "name": "BaseBdev1", 00:09:45.515 "uuid": "471a3f15-cfbf-463d-9206-ae2497843ab5", 00:09:45.515 "is_configured": true, 00:09:45.515 "data_offset": 0, 00:09:45.515 "data_size": 65536 00:09:45.515 }, 00:09:45.515 { 00:09:45.515 "name": null, 00:09:45.515 "uuid": "56be89cb-98c9-4fed-b3b8-32dc061d606a", 00:09:45.515 "is_configured": false, 00:09:45.515 "data_offset": 0, 00:09:45.515 "data_size": 65536 00:09:45.515 }, 00:09:45.515 { 00:09:45.515 "name": "BaseBdev3", 00:09:45.515 "uuid": "22b62ba9-36ce-42d7-94d7-85cf99756e33", 00:09:45.515 "is_configured": true, 00:09:45.515 "data_offset": 0, 00:09:45.515 "data_size": 65536 00:09:45.515 }, 00:09:45.515 { 00:09:45.515 "name": "BaseBdev4", 00:09:45.515 "uuid": "fb647cf8-0896-4972-b9c9-806e17972047", 00:09:45.515 "is_configured": true, 00:09:45.515 "data_offset": 0, 00:09:45.515 "data_size": 65536 00:09:45.515 } 00:09:45.515 ] 00:09:45.515 }' 00:09:45.516 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:45.516 15:14:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.775 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:45.775 15:14:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.775 15:14:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.775 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:45.775 15:14:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.775 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:45.775 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:45.775 15:14:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.775 15:14:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.775 [2024-11-27 15:14:13.790209] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:45.775 15:14:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.775 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:45.775 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:45.775 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:45.775 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:45.775 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:45.775 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:45.775 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:45.775 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:45.775 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:45.775 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:45.775 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:45.775 15:14:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.775 15:14:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.775 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:45.775 15:14:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.775 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:45.775 "name": "Existed_Raid", 00:09:45.775 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:45.775 "strip_size_kb": 64, 00:09:45.775 "state": "configuring", 00:09:45.775 "raid_level": "raid0", 00:09:45.775 "superblock": false, 00:09:45.775 "num_base_bdevs": 4, 00:09:45.775 "num_base_bdevs_discovered": 2, 00:09:45.775 "num_base_bdevs_operational": 4, 00:09:45.775 "base_bdevs_list": [ 00:09:45.775 { 00:09:45.775 "name": null, 00:09:45.775 "uuid": "471a3f15-cfbf-463d-9206-ae2497843ab5", 00:09:45.775 "is_configured": false, 00:09:45.775 "data_offset": 0, 00:09:45.775 "data_size": 65536 00:09:45.775 }, 00:09:45.775 { 00:09:45.775 "name": null, 00:09:45.775 "uuid": "56be89cb-98c9-4fed-b3b8-32dc061d606a", 00:09:45.775 "is_configured": false, 00:09:45.775 "data_offset": 0, 00:09:45.775 "data_size": 65536 00:09:45.775 }, 00:09:45.775 { 00:09:45.775 "name": "BaseBdev3", 00:09:45.775 "uuid": "22b62ba9-36ce-42d7-94d7-85cf99756e33", 00:09:45.775 "is_configured": true, 00:09:45.775 "data_offset": 0, 00:09:45.775 "data_size": 65536 00:09:45.775 }, 00:09:45.775 { 00:09:45.775 "name": "BaseBdev4", 00:09:45.775 "uuid": "fb647cf8-0896-4972-b9c9-806e17972047", 00:09:45.775 "is_configured": true, 00:09:45.775 "data_offset": 0, 00:09:45.775 "data_size": 65536 00:09:45.775 } 00:09:45.775 ] 00:09:45.775 }' 00:09:45.775 15:14:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:45.775 15:14:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.343 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:46.343 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:46.343 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.343 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.343 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.343 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:46.343 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:46.343 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.343 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.343 [2024-11-27 15:14:14.291823] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:46.343 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.343 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:46.343 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:46.343 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:46.343 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:46.343 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:46.343 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:46.343 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:46.343 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:46.343 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:46.343 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:46.343 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:46.343 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:46.343 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.343 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.343 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.343 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:46.343 "name": "Existed_Raid", 00:09:46.343 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:46.343 "strip_size_kb": 64, 00:09:46.343 "state": "configuring", 00:09:46.343 "raid_level": "raid0", 00:09:46.343 "superblock": false, 00:09:46.343 "num_base_bdevs": 4, 00:09:46.343 "num_base_bdevs_discovered": 3, 00:09:46.343 "num_base_bdevs_operational": 4, 00:09:46.343 "base_bdevs_list": [ 00:09:46.343 { 00:09:46.343 "name": null, 00:09:46.343 "uuid": "471a3f15-cfbf-463d-9206-ae2497843ab5", 00:09:46.343 "is_configured": false, 00:09:46.343 "data_offset": 0, 00:09:46.343 "data_size": 65536 00:09:46.343 }, 00:09:46.343 { 00:09:46.343 "name": "BaseBdev2", 00:09:46.343 "uuid": "56be89cb-98c9-4fed-b3b8-32dc061d606a", 00:09:46.343 "is_configured": true, 00:09:46.343 "data_offset": 0, 00:09:46.343 "data_size": 65536 00:09:46.343 }, 00:09:46.343 { 00:09:46.343 "name": "BaseBdev3", 00:09:46.343 "uuid": "22b62ba9-36ce-42d7-94d7-85cf99756e33", 00:09:46.343 "is_configured": true, 00:09:46.344 "data_offset": 0, 00:09:46.344 "data_size": 65536 00:09:46.344 }, 00:09:46.344 { 00:09:46.344 "name": "BaseBdev4", 00:09:46.344 "uuid": "fb647cf8-0896-4972-b9c9-806e17972047", 00:09:46.344 "is_configured": true, 00:09:46.344 "data_offset": 0, 00:09:46.344 "data_size": 65536 00:09:46.344 } 00:09:46.344 ] 00:09:46.344 }' 00:09:46.344 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:46.344 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.912 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:46.912 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.912 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.912 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:46.912 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.912 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:46.912 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:46.912 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:46.912 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.912 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.912 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.912 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 471a3f15-cfbf-463d-9206-ae2497843ab5 00:09:46.912 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.912 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.912 [2024-11-27 15:14:14.857880] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:46.912 [2024-11-27 15:14:14.857997] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:09:46.912 [2024-11-27 15:14:14.858022] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:09:46.912 [2024-11-27 15:14:14.858310] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:09:46.912 [2024-11-27 15:14:14.858462] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:09:46.912 [2024-11-27 15:14:14.858503] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006d00 00:09:46.912 [2024-11-27 15:14:14.858706] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:46.912 NewBaseBdev 00:09:46.912 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.912 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:46.912 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:09:46.912 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:46.912 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:46.912 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:46.912 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:46.912 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:46.912 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.912 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.912 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.912 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:46.912 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.912 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.912 [ 00:09:46.912 { 00:09:46.912 "name": "NewBaseBdev", 00:09:46.912 "aliases": [ 00:09:46.912 "471a3f15-cfbf-463d-9206-ae2497843ab5" 00:09:46.913 ], 00:09:46.913 "product_name": "Malloc disk", 00:09:46.913 "block_size": 512, 00:09:46.913 "num_blocks": 65536, 00:09:46.913 "uuid": "471a3f15-cfbf-463d-9206-ae2497843ab5", 00:09:46.913 "assigned_rate_limits": { 00:09:46.913 "rw_ios_per_sec": 0, 00:09:46.913 "rw_mbytes_per_sec": 0, 00:09:46.913 "r_mbytes_per_sec": 0, 00:09:46.913 "w_mbytes_per_sec": 0 00:09:46.913 }, 00:09:46.913 "claimed": true, 00:09:46.913 "claim_type": "exclusive_write", 00:09:46.913 "zoned": false, 00:09:46.913 "supported_io_types": { 00:09:46.913 "read": true, 00:09:46.913 "write": true, 00:09:46.913 "unmap": true, 00:09:46.913 "flush": true, 00:09:46.913 "reset": true, 00:09:46.913 "nvme_admin": false, 00:09:46.913 "nvme_io": false, 00:09:46.913 "nvme_io_md": false, 00:09:46.913 "write_zeroes": true, 00:09:46.913 "zcopy": true, 00:09:46.913 "get_zone_info": false, 00:09:46.913 "zone_management": false, 00:09:46.913 "zone_append": false, 00:09:46.913 "compare": false, 00:09:46.913 "compare_and_write": false, 00:09:46.913 "abort": true, 00:09:46.913 "seek_hole": false, 00:09:46.913 "seek_data": false, 00:09:46.913 "copy": true, 00:09:46.913 "nvme_iov_md": false 00:09:46.913 }, 00:09:46.913 "memory_domains": [ 00:09:46.913 { 00:09:46.913 "dma_device_id": "system", 00:09:46.913 "dma_device_type": 1 00:09:46.913 }, 00:09:46.913 { 00:09:46.913 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:46.913 "dma_device_type": 2 00:09:46.913 } 00:09:46.913 ], 00:09:46.913 "driver_specific": {} 00:09:46.913 } 00:09:46.913 ] 00:09:46.913 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.913 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:46.913 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:09:46.913 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:46.913 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:46.913 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:46.913 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:46.913 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:46.913 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:46.913 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:46.913 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:46.913 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:46.913 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:46.913 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.913 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.913 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:46.913 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.913 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:46.913 "name": "Existed_Raid", 00:09:46.913 "uuid": "45f20f5e-b3a7-431a-9bf9-6f18954cfc00", 00:09:46.913 "strip_size_kb": 64, 00:09:46.913 "state": "online", 00:09:46.913 "raid_level": "raid0", 00:09:46.913 "superblock": false, 00:09:46.913 "num_base_bdevs": 4, 00:09:46.913 "num_base_bdevs_discovered": 4, 00:09:46.913 "num_base_bdevs_operational": 4, 00:09:46.913 "base_bdevs_list": [ 00:09:46.913 { 00:09:46.913 "name": "NewBaseBdev", 00:09:46.913 "uuid": "471a3f15-cfbf-463d-9206-ae2497843ab5", 00:09:46.913 "is_configured": true, 00:09:46.913 "data_offset": 0, 00:09:46.913 "data_size": 65536 00:09:46.913 }, 00:09:46.913 { 00:09:46.913 "name": "BaseBdev2", 00:09:46.913 "uuid": "56be89cb-98c9-4fed-b3b8-32dc061d606a", 00:09:46.913 "is_configured": true, 00:09:46.913 "data_offset": 0, 00:09:46.913 "data_size": 65536 00:09:46.913 }, 00:09:46.913 { 00:09:46.913 "name": "BaseBdev3", 00:09:46.913 "uuid": "22b62ba9-36ce-42d7-94d7-85cf99756e33", 00:09:46.913 "is_configured": true, 00:09:46.913 "data_offset": 0, 00:09:46.913 "data_size": 65536 00:09:46.913 }, 00:09:46.913 { 00:09:46.913 "name": "BaseBdev4", 00:09:46.913 "uuid": "fb647cf8-0896-4972-b9c9-806e17972047", 00:09:46.913 "is_configured": true, 00:09:46.913 "data_offset": 0, 00:09:46.913 "data_size": 65536 00:09:46.913 } 00:09:46.913 ] 00:09:46.913 }' 00:09:46.913 15:14:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:46.913 15:14:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.483 [2024-11-27 15:14:15.313502] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:47.483 "name": "Existed_Raid", 00:09:47.483 "aliases": [ 00:09:47.483 "45f20f5e-b3a7-431a-9bf9-6f18954cfc00" 00:09:47.483 ], 00:09:47.483 "product_name": "Raid Volume", 00:09:47.483 "block_size": 512, 00:09:47.483 "num_blocks": 262144, 00:09:47.483 "uuid": "45f20f5e-b3a7-431a-9bf9-6f18954cfc00", 00:09:47.483 "assigned_rate_limits": { 00:09:47.483 "rw_ios_per_sec": 0, 00:09:47.483 "rw_mbytes_per_sec": 0, 00:09:47.483 "r_mbytes_per_sec": 0, 00:09:47.483 "w_mbytes_per_sec": 0 00:09:47.483 }, 00:09:47.483 "claimed": false, 00:09:47.483 "zoned": false, 00:09:47.483 "supported_io_types": { 00:09:47.483 "read": true, 00:09:47.483 "write": true, 00:09:47.483 "unmap": true, 00:09:47.483 "flush": true, 00:09:47.483 "reset": true, 00:09:47.483 "nvme_admin": false, 00:09:47.483 "nvme_io": false, 00:09:47.483 "nvme_io_md": false, 00:09:47.483 "write_zeroes": true, 00:09:47.483 "zcopy": false, 00:09:47.483 "get_zone_info": false, 00:09:47.483 "zone_management": false, 00:09:47.483 "zone_append": false, 00:09:47.483 "compare": false, 00:09:47.483 "compare_and_write": false, 00:09:47.483 "abort": false, 00:09:47.483 "seek_hole": false, 00:09:47.483 "seek_data": false, 00:09:47.483 "copy": false, 00:09:47.483 "nvme_iov_md": false 00:09:47.483 }, 00:09:47.483 "memory_domains": [ 00:09:47.483 { 00:09:47.483 "dma_device_id": "system", 00:09:47.483 "dma_device_type": 1 00:09:47.483 }, 00:09:47.483 { 00:09:47.483 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:47.483 "dma_device_type": 2 00:09:47.483 }, 00:09:47.483 { 00:09:47.483 "dma_device_id": "system", 00:09:47.483 "dma_device_type": 1 00:09:47.483 }, 00:09:47.483 { 00:09:47.483 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:47.483 "dma_device_type": 2 00:09:47.483 }, 00:09:47.483 { 00:09:47.483 "dma_device_id": "system", 00:09:47.483 "dma_device_type": 1 00:09:47.483 }, 00:09:47.483 { 00:09:47.483 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:47.483 "dma_device_type": 2 00:09:47.483 }, 00:09:47.483 { 00:09:47.483 "dma_device_id": "system", 00:09:47.483 "dma_device_type": 1 00:09:47.483 }, 00:09:47.483 { 00:09:47.483 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:47.483 "dma_device_type": 2 00:09:47.483 } 00:09:47.483 ], 00:09:47.483 "driver_specific": { 00:09:47.483 "raid": { 00:09:47.483 "uuid": "45f20f5e-b3a7-431a-9bf9-6f18954cfc00", 00:09:47.483 "strip_size_kb": 64, 00:09:47.483 "state": "online", 00:09:47.483 "raid_level": "raid0", 00:09:47.483 "superblock": false, 00:09:47.483 "num_base_bdevs": 4, 00:09:47.483 "num_base_bdevs_discovered": 4, 00:09:47.483 "num_base_bdevs_operational": 4, 00:09:47.483 "base_bdevs_list": [ 00:09:47.483 { 00:09:47.483 "name": "NewBaseBdev", 00:09:47.483 "uuid": "471a3f15-cfbf-463d-9206-ae2497843ab5", 00:09:47.483 "is_configured": true, 00:09:47.483 "data_offset": 0, 00:09:47.483 "data_size": 65536 00:09:47.483 }, 00:09:47.483 { 00:09:47.483 "name": "BaseBdev2", 00:09:47.483 "uuid": "56be89cb-98c9-4fed-b3b8-32dc061d606a", 00:09:47.483 "is_configured": true, 00:09:47.483 "data_offset": 0, 00:09:47.483 "data_size": 65536 00:09:47.483 }, 00:09:47.483 { 00:09:47.483 "name": "BaseBdev3", 00:09:47.483 "uuid": "22b62ba9-36ce-42d7-94d7-85cf99756e33", 00:09:47.483 "is_configured": true, 00:09:47.483 "data_offset": 0, 00:09:47.483 "data_size": 65536 00:09:47.483 }, 00:09:47.483 { 00:09:47.483 "name": "BaseBdev4", 00:09:47.483 "uuid": "fb647cf8-0896-4972-b9c9-806e17972047", 00:09:47.483 "is_configured": true, 00:09:47.483 "data_offset": 0, 00:09:47.483 "data_size": 65536 00:09:47.483 } 00:09:47.483 ] 00:09:47.483 } 00:09:47.483 } 00:09:47.483 }' 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:47.483 BaseBdev2 00:09:47.483 BaseBdev3 00:09:47.483 BaseBdev4' 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:47.483 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:47.484 15:14:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.484 15:14:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.743 15:14:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.743 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:47.743 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:47.743 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:47.743 15:14:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.744 15:14:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.744 [2024-11-27 15:14:15.620648] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:47.744 [2024-11-27 15:14:15.620726] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:47.744 [2024-11-27 15:14:15.620837] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:47.744 [2024-11-27 15:14:15.620949] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:47.744 [2024-11-27 15:14:15.621019] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name Existed_Raid, state offline 00:09:47.744 15:14:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.744 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 80518 00:09:47.744 15:14:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 80518 ']' 00:09:47.744 15:14:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 80518 00:09:47.744 15:14:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:09:47.744 15:14:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:47.744 15:14:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 80518 00:09:47.744 15:14:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:47.744 15:14:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:47.744 15:14:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 80518' 00:09:47.744 killing process with pid 80518 00:09:47.744 15:14:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 80518 00:09:47.744 [2024-11-27 15:14:15.667547] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:47.744 15:14:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 80518 00:09:47.744 [2024-11-27 15:14:15.709254] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:48.003 15:14:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:09:48.003 00:09:48.003 real 0m9.349s 00:09:48.003 user 0m15.905s 00:09:48.003 sys 0m2.054s 00:09:48.003 ************************************ 00:09:48.003 END TEST raid_state_function_test 00:09:48.003 ************************************ 00:09:48.003 15:14:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:48.003 15:14:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.003 15:14:15 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 4 true 00:09:48.003 15:14:15 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:48.003 15:14:15 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:48.003 15:14:15 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:48.003 ************************************ 00:09:48.003 START TEST raid_state_function_test_sb 00:09:48.003 ************************************ 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 4 true 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=81162 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 81162' 00:09:48.003 Process raid pid: 81162 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 81162 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 81162 ']' 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:48.003 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:48.003 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:48.263 [2024-11-27 15:14:16.108421] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:09:48.263 [2024-11-27 15:14:16.108547] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:48.263 [2024-11-27 15:14:16.278311] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:48.263 [2024-11-27 15:14:16.308529] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:48.263 [2024-11-27 15:14:16.351636] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:48.263 [2024-11-27 15:14:16.351670] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:49.202 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:49.202 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:09:49.202 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:49.202 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.202 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:49.202 [2024-11-27 15:14:16.950941] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:49.202 [2024-11-27 15:14:16.951054] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:49.202 [2024-11-27 15:14:16.951083] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:49.202 [2024-11-27 15:14:16.951107] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:49.202 [2024-11-27 15:14:16.951126] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:49.202 [2024-11-27 15:14:16.951149] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:49.202 [2024-11-27 15:14:16.951166] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:49.202 [2024-11-27 15:14:16.951186] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:49.202 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.202 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:49.202 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:49.202 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:49.202 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:49.202 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:49.202 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:49.202 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:49.202 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:49.202 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:49.202 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:49.202 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:49.202 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:49.202 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.202 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:49.202 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.202 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:49.202 "name": "Existed_Raid", 00:09:49.202 "uuid": "a5102229-feba-4570-8944-9a104ac4d040", 00:09:49.202 "strip_size_kb": 64, 00:09:49.202 "state": "configuring", 00:09:49.202 "raid_level": "raid0", 00:09:49.202 "superblock": true, 00:09:49.202 "num_base_bdevs": 4, 00:09:49.202 "num_base_bdevs_discovered": 0, 00:09:49.202 "num_base_bdevs_operational": 4, 00:09:49.202 "base_bdevs_list": [ 00:09:49.202 { 00:09:49.202 "name": "BaseBdev1", 00:09:49.202 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:49.202 "is_configured": false, 00:09:49.202 "data_offset": 0, 00:09:49.202 "data_size": 0 00:09:49.202 }, 00:09:49.202 { 00:09:49.202 "name": "BaseBdev2", 00:09:49.202 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:49.202 "is_configured": false, 00:09:49.202 "data_offset": 0, 00:09:49.202 "data_size": 0 00:09:49.202 }, 00:09:49.202 { 00:09:49.202 "name": "BaseBdev3", 00:09:49.202 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:49.202 "is_configured": false, 00:09:49.202 "data_offset": 0, 00:09:49.202 "data_size": 0 00:09:49.202 }, 00:09:49.202 { 00:09:49.202 "name": "BaseBdev4", 00:09:49.202 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:49.202 "is_configured": false, 00:09:49.202 "data_offset": 0, 00:09:49.202 "data_size": 0 00:09:49.202 } 00:09:49.202 ] 00:09:49.202 }' 00:09:49.202 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:49.202 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:49.462 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:49.462 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.462 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:49.462 [2024-11-27 15:14:17.366237] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:49.462 [2024-11-27 15:14:17.366378] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:49.463 [2024-11-27 15:14:17.378187] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:49.463 [2024-11-27 15:14:17.378297] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:49.463 [2024-11-27 15:14:17.378324] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:49.463 [2024-11-27 15:14:17.378346] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:49.463 [2024-11-27 15:14:17.378364] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:49.463 [2024-11-27 15:14:17.378384] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:49.463 [2024-11-27 15:14:17.378401] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:49.463 [2024-11-27 15:14:17.378421] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:49.463 [2024-11-27 15:14:17.399223] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:49.463 BaseBdev1 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:49.463 [ 00:09:49.463 { 00:09:49.463 "name": "BaseBdev1", 00:09:49.463 "aliases": [ 00:09:49.463 "18b5efe2-6ae3-4488-9948-b2b1d1522131" 00:09:49.463 ], 00:09:49.463 "product_name": "Malloc disk", 00:09:49.463 "block_size": 512, 00:09:49.463 "num_blocks": 65536, 00:09:49.463 "uuid": "18b5efe2-6ae3-4488-9948-b2b1d1522131", 00:09:49.463 "assigned_rate_limits": { 00:09:49.463 "rw_ios_per_sec": 0, 00:09:49.463 "rw_mbytes_per_sec": 0, 00:09:49.463 "r_mbytes_per_sec": 0, 00:09:49.463 "w_mbytes_per_sec": 0 00:09:49.463 }, 00:09:49.463 "claimed": true, 00:09:49.463 "claim_type": "exclusive_write", 00:09:49.463 "zoned": false, 00:09:49.463 "supported_io_types": { 00:09:49.463 "read": true, 00:09:49.463 "write": true, 00:09:49.463 "unmap": true, 00:09:49.463 "flush": true, 00:09:49.463 "reset": true, 00:09:49.463 "nvme_admin": false, 00:09:49.463 "nvme_io": false, 00:09:49.463 "nvme_io_md": false, 00:09:49.463 "write_zeroes": true, 00:09:49.463 "zcopy": true, 00:09:49.463 "get_zone_info": false, 00:09:49.463 "zone_management": false, 00:09:49.463 "zone_append": false, 00:09:49.463 "compare": false, 00:09:49.463 "compare_and_write": false, 00:09:49.463 "abort": true, 00:09:49.463 "seek_hole": false, 00:09:49.463 "seek_data": false, 00:09:49.463 "copy": true, 00:09:49.463 "nvme_iov_md": false 00:09:49.463 }, 00:09:49.463 "memory_domains": [ 00:09:49.463 { 00:09:49.463 "dma_device_id": "system", 00:09:49.463 "dma_device_type": 1 00:09:49.463 }, 00:09:49.463 { 00:09:49.463 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:49.463 "dma_device_type": 2 00:09:49.463 } 00:09:49.463 ], 00:09:49.463 "driver_specific": {} 00:09:49.463 } 00:09:49.463 ] 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:49.463 "name": "Existed_Raid", 00:09:49.463 "uuid": "357a7f35-5299-4d57-acd5-62b443b5ead2", 00:09:49.463 "strip_size_kb": 64, 00:09:49.463 "state": "configuring", 00:09:49.463 "raid_level": "raid0", 00:09:49.463 "superblock": true, 00:09:49.463 "num_base_bdevs": 4, 00:09:49.463 "num_base_bdevs_discovered": 1, 00:09:49.463 "num_base_bdevs_operational": 4, 00:09:49.463 "base_bdevs_list": [ 00:09:49.463 { 00:09:49.463 "name": "BaseBdev1", 00:09:49.463 "uuid": "18b5efe2-6ae3-4488-9948-b2b1d1522131", 00:09:49.463 "is_configured": true, 00:09:49.463 "data_offset": 2048, 00:09:49.463 "data_size": 63488 00:09:49.463 }, 00:09:49.463 { 00:09:49.463 "name": "BaseBdev2", 00:09:49.463 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:49.463 "is_configured": false, 00:09:49.463 "data_offset": 0, 00:09:49.463 "data_size": 0 00:09:49.463 }, 00:09:49.463 { 00:09:49.463 "name": "BaseBdev3", 00:09:49.463 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:49.463 "is_configured": false, 00:09:49.463 "data_offset": 0, 00:09:49.463 "data_size": 0 00:09:49.463 }, 00:09:49.463 { 00:09:49.463 "name": "BaseBdev4", 00:09:49.463 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:49.463 "is_configured": false, 00:09:49.463 "data_offset": 0, 00:09:49.463 "data_size": 0 00:09:49.463 } 00:09:49.463 ] 00:09:49.463 }' 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:49.463 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.033 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:50.033 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.033 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.033 [2024-11-27 15:14:17.874448] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:50.033 [2024-11-27 15:14:17.874554] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:09:50.033 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.033 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:50.033 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.033 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.033 [2024-11-27 15:14:17.886457] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:50.033 [2024-11-27 15:14:17.888339] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:50.033 [2024-11-27 15:14:17.888424] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:50.033 [2024-11-27 15:14:17.888454] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:50.033 [2024-11-27 15:14:17.888477] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:50.033 [2024-11-27 15:14:17.888495] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:50.033 [2024-11-27 15:14:17.888533] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:50.033 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.033 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:50.033 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:50.033 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:50.033 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:50.033 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:50.033 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:50.033 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:50.033 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:50.033 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:50.033 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:50.034 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:50.034 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:50.034 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:50.034 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.034 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.034 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:50.034 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.034 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:50.034 "name": "Existed_Raid", 00:09:50.034 "uuid": "64e44a91-69b6-4525-8113-3914f25db640", 00:09:50.034 "strip_size_kb": 64, 00:09:50.034 "state": "configuring", 00:09:50.034 "raid_level": "raid0", 00:09:50.034 "superblock": true, 00:09:50.034 "num_base_bdevs": 4, 00:09:50.034 "num_base_bdevs_discovered": 1, 00:09:50.034 "num_base_bdevs_operational": 4, 00:09:50.034 "base_bdevs_list": [ 00:09:50.034 { 00:09:50.034 "name": "BaseBdev1", 00:09:50.034 "uuid": "18b5efe2-6ae3-4488-9948-b2b1d1522131", 00:09:50.034 "is_configured": true, 00:09:50.034 "data_offset": 2048, 00:09:50.034 "data_size": 63488 00:09:50.034 }, 00:09:50.034 { 00:09:50.034 "name": "BaseBdev2", 00:09:50.034 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:50.034 "is_configured": false, 00:09:50.034 "data_offset": 0, 00:09:50.034 "data_size": 0 00:09:50.034 }, 00:09:50.034 { 00:09:50.034 "name": "BaseBdev3", 00:09:50.034 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:50.034 "is_configured": false, 00:09:50.034 "data_offset": 0, 00:09:50.034 "data_size": 0 00:09:50.034 }, 00:09:50.034 { 00:09:50.034 "name": "BaseBdev4", 00:09:50.034 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:50.034 "is_configured": false, 00:09:50.034 "data_offset": 0, 00:09:50.034 "data_size": 0 00:09:50.034 } 00:09:50.034 ] 00:09:50.034 }' 00:09:50.034 15:14:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:50.034 15:14:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.294 [2024-11-27 15:14:18.352637] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:50.294 BaseBdev2 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.294 [ 00:09:50.294 { 00:09:50.294 "name": "BaseBdev2", 00:09:50.294 "aliases": [ 00:09:50.294 "bdaf39a4-cf02-43f1-8ce3-92bbfcb01021" 00:09:50.294 ], 00:09:50.294 "product_name": "Malloc disk", 00:09:50.294 "block_size": 512, 00:09:50.294 "num_blocks": 65536, 00:09:50.294 "uuid": "bdaf39a4-cf02-43f1-8ce3-92bbfcb01021", 00:09:50.294 "assigned_rate_limits": { 00:09:50.294 "rw_ios_per_sec": 0, 00:09:50.294 "rw_mbytes_per_sec": 0, 00:09:50.294 "r_mbytes_per_sec": 0, 00:09:50.294 "w_mbytes_per_sec": 0 00:09:50.294 }, 00:09:50.294 "claimed": true, 00:09:50.294 "claim_type": "exclusive_write", 00:09:50.294 "zoned": false, 00:09:50.294 "supported_io_types": { 00:09:50.294 "read": true, 00:09:50.294 "write": true, 00:09:50.294 "unmap": true, 00:09:50.294 "flush": true, 00:09:50.294 "reset": true, 00:09:50.294 "nvme_admin": false, 00:09:50.294 "nvme_io": false, 00:09:50.294 "nvme_io_md": false, 00:09:50.294 "write_zeroes": true, 00:09:50.294 "zcopy": true, 00:09:50.294 "get_zone_info": false, 00:09:50.294 "zone_management": false, 00:09:50.294 "zone_append": false, 00:09:50.294 "compare": false, 00:09:50.294 "compare_and_write": false, 00:09:50.294 "abort": true, 00:09:50.294 "seek_hole": false, 00:09:50.294 "seek_data": false, 00:09:50.294 "copy": true, 00:09:50.294 "nvme_iov_md": false 00:09:50.294 }, 00:09:50.294 "memory_domains": [ 00:09:50.294 { 00:09:50.294 "dma_device_id": "system", 00:09:50.294 "dma_device_type": 1 00:09:50.294 }, 00:09:50.294 { 00:09:50.294 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:50.294 "dma_device_type": 2 00:09:50.294 } 00:09:50.294 ], 00:09:50.294 "driver_specific": {} 00:09:50.294 } 00:09:50.294 ] 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:50.294 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.554 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.554 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.554 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:50.554 "name": "Existed_Raid", 00:09:50.554 "uuid": "64e44a91-69b6-4525-8113-3914f25db640", 00:09:50.554 "strip_size_kb": 64, 00:09:50.554 "state": "configuring", 00:09:50.554 "raid_level": "raid0", 00:09:50.554 "superblock": true, 00:09:50.554 "num_base_bdevs": 4, 00:09:50.554 "num_base_bdevs_discovered": 2, 00:09:50.554 "num_base_bdevs_operational": 4, 00:09:50.554 "base_bdevs_list": [ 00:09:50.554 { 00:09:50.554 "name": "BaseBdev1", 00:09:50.554 "uuid": "18b5efe2-6ae3-4488-9948-b2b1d1522131", 00:09:50.554 "is_configured": true, 00:09:50.554 "data_offset": 2048, 00:09:50.554 "data_size": 63488 00:09:50.554 }, 00:09:50.554 { 00:09:50.554 "name": "BaseBdev2", 00:09:50.554 "uuid": "bdaf39a4-cf02-43f1-8ce3-92bbfcb01021", 00:09:50.554 "is_configured": true, 00:09:50.554 "data_offset": 2048, 00:09:50.554 "data_size": 63488 00:09:50.554 }, 00:09:50.554 { 00:09:50.554 "name": "BaseBdev3", 00:09:50.554 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:50.554 "is_configured": false, 00:09:50.554 "data_offset": 0, 00:09:50.554 "data_size": 0 00:09:50.554 }, 00:09:50.554 { 00:09:50.554 "name": "BaseBdev4", 00:09:50.554 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:50.554 "is_configured": false, 00:09:50.554 "data_offset": 0, 00:09:50.554 "data_size": 0 00:09:50.554 } 00:09:50.554 ] 00:09:50.554 }' 00:09:50.554 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:50.554 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.828 [2024-11-27 15:14:18.846174] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:50.828 BaseBdev3 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.828 [ 00:09:50.828 { 00:09:50.828 "name": "BaseBdev3", 00:09:50.828 "aliases": [ 00:09:50.828 "864e6d57-c7f0-4ec6-9e8e-bc58664245a6" 00:09:50.828 ], 00:09:50.828 "product_name": "Malloc disk", 00:09:50.828 "block_size": 512, 00:09:50.828 "num_blocks": 65536, 00:09:50.828 "uuid": "864e6d57-c7f0-4ec6-9e8e-bc58664245a6", 00:09:50.828 "assigned_rate_limits": { 00:09:50.828 "rw_ios_per_sec": 0, 00:09:50.828 "rw_mbytes_per_sec": 0, 00:09:50.828 "r_mbytes_per_sec": 0, 00:09:50.828 "w_mbytes_per_sec": 0 00:09:50.828 }, 00:09:50.828 "claimed": true, 00:09:50.828 "claim_type": "exclusive_write", 00:09:50.828 "zoned": false, 00:09:50.828 "supported_io_types": { 00:09:50.828 "read": true, 00:09:50.828 "write": true, 00:09:50.828 "unmap": true, 00:09:50.828 "flush": true, 00:09:50.828 "reset": true, 00:09:50.828 "nvme_admin": false, 00:09:50.828 "nvme_io": false, 00:09:50.828 "nvme_io_md": false, 00:09:50.828 "write_zeroes": true, 00:09:50.828 "zcopy": true, 00:09:50.828 "get_zone_info": false, 00:09:50.828 "zone_management": false, 00:09:50.828 "zone_append": false, 00:09:50.828 "compare": false, 00:09:50.828 "compare_and_write": false, 00:09:50.828 "abort": true, 00:09:50.828 "seek_hole": false, 00:09:50.828 "seek_data": false, 00:09:50.828 "copy": true, 00:09:50.828 "nvme_iov_md": false 00:09:50.828 }, 00:09:50.828 "memory_domains": [ 00:09:50.828 { 00:09:50.828 "dma_device_id": "system", 00:09:50.828 "dma_device_type": 1 00:09:50.828 }, 00:09:50.828 { 00:09:50.828 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:50.828 "dma_device_type": 2 00:09:50.828 } 00:09:50.828 ], 00:09:50.828 "driver_specific": {} 00:09:50.828 } 00:09:50.828 ] 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.828 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.089 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:51.089 "name": "Existed_Raid", 00:09:51.089 "uuid": "64e44a91-69b6-4525-8113-3914f25db640", 00:09:51.089 "strip_size_kb": 64, 00:09:51.089 "state": "configuring", 00:09:51.089 "raid_level": "raid0", 00:09:51.089 "superblock": true, 00:09:51.089 "num_base_bdevs": 4, 00:09:51.089 "num_base_bdevs_discovered": 3, 00:09:51.089 "num_base_bdevs_operational": 4, 00:09:51.089 "base_bdevs_list": [ 00:09:51.089 { 00:09:51.089 "name": "BaseBdev1", 00:09:51.089 "uuid": "18b5efe2-6ae3-4488-9948-b2b1d1522131", 00:09:51.089 "is_configured": true, 00:09:51.089 "data_offset": 2048, 00:09:51.089 "data_size": 63488 00:09:51.089 }, 00:09:51.089 { 00:09:51.089 "name": "BaseBdev2", 00:09:51.089 "uuid": "bdaf39a4-cf02-43f1-8ce3-92bbfcb01021", 00:09:51.089 "is_configured": true, 00:09:51.089 "data_offset": 2048, 00:09:51.089 "data_size": 63488 00:09:51.089 }, 00:09:51.089 { 00:09:51.089 "name": "BaseBdev3", 00:09:51.089 "uuid": "864e6d57-c7f0-4ec6-9e8e-bc58664245a6", 00:09:51.089 "is_configured": true, 00:09:51.089 "data_offset": 2048, 00:09:51.089 "data_size": 63488 00:09:51.089 }, 00:09:51.089 { 00:09:51.089 "name": "BaseBdev4", 00:09:51.089 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:51.089 "is_configured": false, 00:09:51.089 "data_offset": 0, 00:09:51.089 "data_size": 0 00:09:51.089 } 00:09:51.089 ] 00:09:51.089 }' 00:09:51.089 15:14:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:51.089 15:14:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:51.349 [2024-11-27 15:14:19.332526] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:51.349 [2024-11-27 15:14:19.332827] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:09:51.349 [2024-11-27 15:14:19.332865] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:51.349 [2024-11-27 15:14:19.333214] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:09:51.349 BaseBdev4 00:09:51.349 [2024-11-27 15:14:19.333383] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:09:51.349 [2024-11-27 15:14:19.333426] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:09:51.349 [2024-11-27 15:14:19.333594] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:51.349 [ 00:09:51.349 { 00:09:51.349 "name": "BaseBdev4", 00:09:51.349 "aliases": [ 00:09:51.349 "32ac7270-d0fd-477a-a921-367596453e7f" 00:09:51.349 ], 00:09:51.349 "product_name": "Malloc disk", 00:09:51.349 "block_size": 512, 00:09:51.349 "num_blocks": 65536, 00:09:51.349 "uuid": "32ac7270-d0fd-477a-a921-367596453e7f", 00:09:51.349 "assigned_rate_limits": { 00:09:51.349 "rw_ios_per_sec": 0, 00:09:51.349 "rw_mbytes_per_sec": 0, 00:09:51.349 "r_mbytes_per_sec": 0, 00:09:51.349 "w_mbytes_per_sec": 0 00:09:51.349 }, 00:09:51.349 "claimed": true, 00:09:51.349 "claim_type": "exclusive_write", 00:09:51.349 "zoned": false, 00:09:51.349 "supported_io_types": { 00:09:51.349 "read": true, 00:09:51.349 "write": true, 00:09:51.349 "unmap": true, 00:09:51.349 "flush": true, 00:09:51.349 "reset": true, 00:09:51.349 "nvme_admin": false, 00:09:51.349 "nvme_io": false, 00:09:51.349 "nvme_io_md": false, 00:09:51.349 "write_zeroes": true, 00:09:51.349 "zcopy": true, 00:09:51.349 "get_zone_info": false, 00:09:51.349 "zone_management": false, 00:09:51.349 "zone_append": false, 00:09:51.349 "compare": false, 00:09:51.349 "compare_and_write": false, 00:09:51.349 "abort": true, 00:09:51.349 "seek_hole": false, 00:09:51.349 "seek_data": false, 00:09:51.349 "copy": true, 00:09:51.349 "nvme_iov_md": false 00:09:51.349 }, 00:09:51.349 "memory_domains": [ 00:09:51.349 { 00:09:51.349 "dma_device_id": "system", 00:09:51.349 "dma_device_type": 1 00:09:51.349 }, 00:09:51.349 { 00:09:51.349 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:51.349 "dma_device_type": 2 00:09:51.349 } 00:09:51.349 ], 00:09:51.349 "driver_specific": {} 00:09:51.349 } 00:09:51.349 ] 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:51.349 "name": "Existed_Raid", 00:09:51.349 "uuid": "64e44a91-69b6-4525-8113-3914f25db640", 00:09:51.349 "strip_size_kb": 64, 00:09:51.349 "state": "online", 00:09:51.349 "raid_level": "raid0", 00:09:51.349 "superblock": true, 00:09:51.349 "num_base_bdevs": 4, 00:09:51.349 "num_base_bdevs_discovered": 4, 00:09:51.349 "num_base_bdevs_operational": 4, 00:09:51.349 "base_bdevs_list": [ 00:09:51.349 { 00:09:51.349 "name": "BaseBdev1", 00:09:51.349 "uuid": "18b5efe2-6ae3-4488-9948-b2b1d1522131", 00:09:51.349 "is_configured": true, 00:09:51.349 "data_offset": 2048, 00:09:51.349 "data_size": 63488 00:09:51.349 }, 00:09:51.349 { 00:09:51.349 "name": "BaseBdev2", 00:09:51.349 "uuid": "bdaf39a4-cf02-43f1-8ce3-92bbfcb01021", 00:09:51.349 "is_configured": true, 00:09:51.349 "data_offset": 2048, 00:09:51.349 "data_size": 63488 00:09:51.349 }, 00:09:51.349 { 00:09:51.349 "name": "BaseBdev3", 00:09:51.349 "uuid": "864e6d57-c7f0-4ec6-9e8e-bc58664245a6", 00:09:51.349 "is_configured": true, 00:09:51.349 "data_offset": 2048, 00:09:51.349 "data_size": 63488 00:09:51.349 }, 00:09:51.349 { 00:09:51.349 "name": "BaseBdev4", 00:09:51.349 "uuid": "32ac7270-d0fd-477a-a921-367596453e7f", 00:09:51.349 "is_configured": true, 00:09:51.349 "data_offset": 2048, 00:09:51.349 "data_size": 63488 00:09:51.349 } 00:09:51.349 ] 00:09:51.349 }' 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:51.349 15:14:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:51.918 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:51.918 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:51.918 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:51.918 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:51.918 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:51.918 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:51.918 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:51.918 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:51.918 15:14:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.918 15:14:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:51.918 [2024-11-27 15:14:19.784202] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:51.918 15:14:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.918 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:51.918 "name": "Existed_Raid", 00:09:51.918 "aliases": [ 00:09:51.918 "64e44a91-69b6-4525-8113-3914f25db640" 00:09:51.918 ], 00:09:51.918 "product_name": "Raid Volume", 00:09:51.918 "block_size": 512, 00:09:51.918 "num_blocks": 253952, 00:09:51.918 "uuid": "64e44a91-69b6-4525-8113-3914f25db640", 00:09:51.918 "assigned_rate_limits": { 00:09:51.918 "rw_ios_per_sec": 0, 00:09:51.918 "rw_mbytes_per_sec": 0, 00:09:51.918 "r_mbytes_per_sec": 0, 00:09:51.918 "w_mbytes_per_sec": 0 00:09:51.918 }, 00:09:51.918 "claimed": false, 00:09:51.918 "zoned": false, 00:09:51.918 "supported_io_types": { 00:09:51.918 "read": true, 00:09:51.918 "write": true, 00:09:51.918 "unmap": true, 00:09:51.918 "flush": true, 00:09:51.918 "reset": true, 00:09:51.918 "nvme_admin": false, 00:09:51.918 "nvme_io": false, 00:09:51.918 "nvme_io_md": false, 00:09:51.918 "write_zeroes": true, 00:09:51.918 "zcopy": false, 00:09:51.918 "get_zone_info": false, 00:09:51.918 "zone_management": false, 00:09:51.918 "zone_append": false, 00:09:51.918 "compare": false, 00:09:51.918 "compare_and_write": false, 00:09:51.918 "abort": false, 00:09:51.918 "seek_hole": false, 00:09:51.918 "seek_data": false, 00:09:51.918 "copy": false, 00:09:51.918 "nvme_iov_md": false 00:09:51.918 }, 00:09:51.918 "memory_domains": [ 00:09:51.918 { 00:09:51.918 "dma_device_id": "system", 00:09:51.918 "dma_device_type": 1 00:09:51.918 }, 00:09:51.918 { 00:09:51.918 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:51.918 "dma_device_type": 2 00:09:51.918 }, 00:09:51.918 { 00:09:51.918 "dma_device_id": "system", 00:09:51.918 "dma_device_type": 1 00:09:51.918 }, 00:09:51.918 { 00:09:51.918 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:51.918 "dma_device_type": 2 00:09:51.918 }, 00:09:51.918 { 00:09:51.918 "dma_device_id": "system", 00:09:51.918 "dma_device_type": 1 00:09:51.918 }, 00:09:51.918 { 00:09:51.918 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:51.918 "dma_device_type": 2 00:09:51.918 }, 00:09:51.918 { 00:09:51.918 "dma_device_id": "system", 00:09:51.918 "dma_device_type": 1 00:09:51.918 }, 00:09:51.918 { 00:09:51.918 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:51.918 "dma_device_type": 2 00:09:51.918 } 00:09:51.918 ], 00:09:51.918 "driver_specific": { 00:09:51.918 "raid": { 00:09:51.918 "uuid": "64e44a91-69b6-4525-8113-3914f25db640", 00:09:51.919 "strip_size_kb": 64, 00:09:51.919 "state": "online", 00:09:51.919 "raid_level": "raid0", 00:09:51.919 "superblock": true, 00:09:51.919 "num_base_bdevs": 4, 00:09:51.919 "num_base_bdevs_discovered": 4, 00:09:51.919 "num_base_bdevs_operational": 4, 00:09:51.919 "base_bdevs_list": [ 00:09:51.919 { 00:09:51.919 "name": "BaseBdev1", 00:09:51.919 "uuid": "18b5efe2-6ae3-4488-9948-b2b1d1522131", 00:09:51.919 "is_configured": true, 00:09:51.919 "data_offset": 2048, 00:09:51.919 "data_size": 63488 00:09:51.919 }, 00:09:51.919 { 00:09:51.919 "name": "BaseBdev2", 00:09:51.919 "uuid": "bdaf39a4-cf02-43f1-8ce3-92bbfcb01021", 00:09:51.919 "is_configured": true, 00:09:51.919 "data_offset": 2048, 00:09:51.919 "data_size": 63488 00:09:51.919 }, 00:09:51.919 { 00:09:51.919 "name": "BaseBdev3", 00:09:51.919 "uuid": "864e6d57-c7f0-4ec6-9e8e-bc58664245a6", 00:09:51.919 "is_configured": true, 00:09:51.919 "data_offset": 2048, 00:09:51.919 "data_size": 63488 00:09:51.919 }, 00:09:51.919 { 00:09:51.919 "name": "BaseBdev4", 00:09:51.919 "uuid": "32ac7270-d0fd-477a-a921-367596453e7f", 00:09:51.919 "is_configured": true, 00:09:51.919 "data_offset": 2048, 00:09:51.919 "data_size": 63488 00:09:51.919 } 00:09:51.919 ] 00:09:51.919 } 00:09:51.919 } 00:09:51.919 }' 00:09:51.919 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:51.919 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:51.919 BaseBdev2 00:09:51.919 BaseBdev3 00:09:51.919 BaseBdev4' 00:09:51.919 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:51.919 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:51.919 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:51.919 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:51.919 15:14:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.919 15:14:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:51.919 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:51.919 15:14:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.919 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:51.919 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:51.919 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:51.919 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:51.919 15:14:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:51.919 15:14:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.919 15:14:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:51.919 15:14:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.919 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:51.919 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:51.919 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:51.919 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:51.919 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:51.919 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.919 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:52.179 [2024-11-27 15:14:20.095331] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:52.179 [2024-11-27 15:14:20.095399] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:52.179 [2024-11-27 15:14:20.095480] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 3 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.179 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:52.179 "name": "Existed_Raid", 00:09:52.179 "uuid": "64e44a91-69b6-4525-8113-3914f25db640", 00:09:52.179 "strip_size_kb": 64, 00:09:52.179 "state": "offline", 00:09:52.179 "raid_level": "raid0", 00:09:52.179 "superblock": true, 00:09:52.179 "num_base_bdevs": 4, 00:09:52.179 "num_base_bdevs_discovered": 3, 00:09:52.179 "num_base_bdevs_operational": 3, 00:09:52.179 "base_bdevs_list": [ 00:09:52.179 { 00:09:52.179 "name": null, 00:09:52.179 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:52.179 "is_configured": false, 00:09:52.179 "data_offset": 0, 00:09:52.179 "data_size": 63488 00:09:52.179 }, 00:09:52.179 { 00:09:52.179 "name": "BaseBdev2", 00:09:52.179 "uuid": "bdaf39a4-cf02-43f1-8ce3-92bbfcb01021", 00:09:52.179 "is_configured": true, 00:09:52.179 "data_offset": 2048, 00:09:52.179 "data_size": 63488 00:09:52.179 }, 00:09:52.179 { 00:09:52.179 "name": "BaseBdev3", 00:09:52.179 "uuid": "864e6d57-c7f0-4ec6-9e8e-bc58664245a6", 00:09:52.179 "is_configured": true, 00:09:52.179 "data_offset": 2048, 00:09:52.179 "data_size": 63488 00:09:52.179 }, 00:09:52.179 { 00:09:52.179 "name": "BaseBdev4", 00:09:52.180 "uuid": "32ac7270-d0fd-477a-a921-367596453e7f", 00:09:52.180 "is_configured": true, 00:09:52.180 "data_offset": 2048, 00:09:52.180 "data_size": 63488 00:09:52.180 } 00:09:52.180 ] 00:09:52.180 }' 00:09:52.180 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:52.180 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:52.439 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:52.439 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:52.439 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:52.439 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:52.700 [2024-11-27 15:14:20.605749] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:52.700 [2024-11-27 15:14:20.676844] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:52.700 [2024-11-27 15:14:20.748040] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:09:52.700 [2024-11-27 15:14:20.748138] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:52.700 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.961 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:52.961 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:52.961 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:09:52.961 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:52.961 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:52.961 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:52.961 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.961 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:52.961 BaseBdev2 00:09:52.961 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.961 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:52.961 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:52.961 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:52.961 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:52.961 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:52.961 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:52.961 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:52.961 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.961 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:52.961 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.961 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:52.961 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.961 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:52.961 [ 00:09:52.961 { 00:09:52.961 "name": "BaseBdev2", 00:09:52.961 "aliases": [ 00:09:52.961 "21af45cc-9c06-414c-b0ac-08ef2a4b44c9" 00:09:52.961 ], 00:09:52.961 "product_name": "Malloc disk", 00:09:52.961 "block_size": 512, 00:09:52.961 "num_blocks": 65536, 00:09:52.961 "uuid": "21af45cc-9c06-414c-b0ac-08ef2a4b44c9", 00:09:52.961 "assigned_rate_limits": { 00:09:52.961 "rw_ios_per_sec": 0, 00:09:52.961 "rw_mbytes_per_sec": 0, 00:09:52.961 "r_mbytes_per_sec": 0, 00:09:52.961 "w_mbytes_per_sec": 0 00:09:52.961 }, 00:09:52.961 "claimed": false, 00:09:52.961 "zoned": false, 00:09:52.961 "supported_io_types": { 00:09:52.961 "read": true, 00:09:52.961 "write": true, 00:09:52.961 "unmap": true, 00:09:52.961 "flush": true, 00:09:52.961 "reset": true, 00:09:52.961 "nvme_admin": false, 00:09:52.961 "nvme_io": false, 00:09:52.961 "nvme_io_md": false, 00:09:52.961 "write_zeroes": true, 00:09:52.961 "zcopy": true, 00:09:52.961 "get_zone_info": false, 00:09:52.961 "zone_management": false, 00:09:52.961 "zone_append": false, 00:09:52.961 "compare": false, 00:09:52.961 "compare_and_write": false, 00:09:52.961 "abort": true, 00:09:52.961 "seek_hole": false, 00:09:52.961 "seek_data": false, 00:09:52.961 "copy": true, 00:09:52.961 "nvme_iov_md": false 00:09:52.961 }, 00:09:52.961 "memory_domains": [ 00:09:52.962 { 00:09:52.962 "dma_device_id": "system", 00:09:52.962 "dma_device_type": 1 00:09:52.962 }, 00:09:52.962 { 00:09:52.962 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:52.962 "dma_device_type": 2 00:09:52.962 } 00:09:52.962 ], 00:09:52.962 "driver_specific": {} 00:09:52.962 } 00:09:52.962 ] 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:52.962 BaseBdev3 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:52.962 [ 00:09:52.962 { 00:09:52.962 "name": "BaseBdev3", 00:09:52.962 "aliases": [ 00:09:52.962 "73e8787e-87c5-4d53-ab2d-570781151829" 00:09:52.962 ], 00:09:52.962 "product_name": "Malloc disk", 00:09:52.962 "block_size": 512, 00:09:52.962 "num_blocks": 65536, 00:09:52.962 "uuid": "73e8787e-87c5-4d53-ab2d-570781151829", 00:09:52.962 "assigned_rate_limits": { 00:09:52.962 "rw_ios_per_sec": 0, 00:09:52.962 "rw_mbytes_per_sec": 0, 00:09:52.962 "r_mbytes_per_sec": 0, 00:09:52.962 "w_mbytes_per_sec": 0 00:09:52.962 }, 00:09:52.962 "claimed": false, 00:09:52.962 "zoned": false, 00:09:52.962 "supported_io_types": { 00:09:52.962 "read": true, 00:09:52.962 "write": true, 00:09:52.962 "unmap": true, 00:09:52.962 "flush": true, 00:09:52.962 "reset": true, 00:09:52.962 "nvme_admin": false, 00:09:52.962 "nvme_io": false, 00:09:52.962 "nvme_io_md": false, 00:09:52.962 "write_zeroes": true, 00:09:52.962 "zcopy": true, 00:09:52.962 "get_zone_info": false, 00:09:52.962 "zone_management": false, 00:09:52.962 "zone_append": false, 00:09:52.962 "compare": false, 00:09:52.962 "compare_and_write": false, 00:09:52.962 "abort": true, 00:09:52.962 "seek_hole": false, 00:09:52.962 "seek_data": false, 00:09:52.962 "copy": true, 00:09:52.962 "nvme_iov_md": false 00:09:52.962 }, 00:09:52.962 "memory_domains": [ 00:09:52.962 { 00:09:52.962 "dma_device_id": "system", 00:09:52.962 "dma_device_type": 1 00:09:52.962 }, 00:09:52.962 { 00:09:52.962 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:52.962 "dma_device_type": 2 00:09:52.962 } 00:09:52.962 ], 00:09:52.962 "driver_specific": {} 00:09:52.962 } 00:09:52.962 ] 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:52.962 BaseBdev4 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:52.962 [ 00:09:52.962 { 00:09:52.962 "name": "BaseBdev4", 00:09:52.962 "aliases": [ 00:09:52.962 "dcb1126c-b50c-4cc2-938f-fe09899d3da3" 00:09:52.962 ], 00:09:52.962 "product_name": "Malloc disk", 00:09:52.962 "block_size": 512, 00:09:52.962 "num_blocks": 65536, 00:09:52.962 "uuid": "dcb1126c-b50c-4cc2-938f-fe09899d3da3", 00:09:52.962 "assigned_rate_limits": { 00:09:52.962 "rw_ios_per_sec": 0, 00:09:52.962 "rw_mbytes_per_sec": 0, 00:09:52.962 "r_mbytes_per_sec": 0, 00:09:52.962 "w_mbytes_per_sec": 0 00:09:52.962 }, 00:09:52.962 "claimed": false, 00:09:52.962 "zoned": false, 00:09:52.962 "supported_io_types": { 00:09:52.962 "read": true, 00:09:52.962 "write": true, 00:09:52.962 "unmap": true, 00:09:52.962 "flush": true, 00:09:52.962 "reset": true, 00:09:52.962 "nvme_admin": false, 00:09:52.962 "nvme_io": false, 00:09:52.962 "nvme_io_md": false, 00:09:52.962 "write_zeroes": true, 00:09:52.962 "zcopy": true, 00:09:52.962 "get_zone_info": false, 00:09:52.962 "zone_management": false, 00:09:52.962 "zone_append": false, 00:09:52.962 "compare": false, 00:09:52.962 "compare_and_write": false, 00:09:52.962 "abort": true, 00:09:52.962 "seek_hole": false, 00:09:52.962 "seek_data": false, 00:09:52.962 "copy": true, 00:09:52.962 "nvme_iov_md": false 00:09:52.962 }, 00:09:52.962 "memory_domains": [ 00:09:52.962 { 00:09:52.962 "dma_device_id": "system", 00:09:52.962 "dma_device_type": 1 00:09:52.962 }, 00:09:52.962 { 00:09:52.962 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:52.962 "dma_device_type": 2 00:09:52.962 } 00:09:52.962 ], 00:09:52.962 "driver_specific": {} 00:09:52.962 } 00:09:52.962 ] 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:52.962 [2024-11-27 15:14:20.965827] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:52.962 [2024-11-27 15:14:20.965932] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:52.962 [2024-11-27 15:14:20.965997] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:52.962 [2024-11-27 15:14:20.967899] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:52.962 [2024-11-27 15:14:20.968005] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:52.962 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:52.963 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:52.963 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:52.963 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:52.963 15:14:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:52.963 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.963 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:52.963 15:14:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.963 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:52.963 "name": "Existed_Raid", 00:09:52.963 "uuid": "a9c658f1-072f-4d01-9aea-4f38b2138950", 00:09:52.963 "strip_size_kb": 64, 00:09:52.963 "state": "configuring", 00:09:52.963 "raid_level": "raid0", 00:09:52.963 "superblock": true, 00:09:52.963 "num_base_bdevs": 4, 00:09:52.963 "num_base_bdevs_discovered": 3, 00:09:52.963 "num_base_bdevs_operational": 4, 00:09:52.963 "base_bdevs_list": [ 00:09:52.963 { 00:09:52.963 "name": "BaseBdev1", 00:09:52.963 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:52.963 "is_configured": false, 00:09:52.963 "data_offset": 0, 00:09:52.963 "data_size": 0 00:09:52.963 }, 00:09:52.963 { 00:09:52.963 "name": "BaseBdev2", 00:09:52.963 "uuid": "21af45cc-9c06-414c-b0ac-08ef2a4b44c9", 00:09:52.963 "is_configured": true, 00:09:52.963 "data_offset": 2048, 00:09:52.963 "data_size": 63488 00:09:52.963 }, 00:09:52.963 { 00:09:52.963 "name": "BaseBdev3", 00:09:52.963 "uuid": "73e8787e-87c5-4d53-ab2d-570781151829", 00:09:52.963 "is_configured": true, 00:09:52.963 "data_offset": 2048, 00:09:52.963 "data_size": 63488 00:09:52.963 }, 00:09:52.963 { 00:09:52.963 "name": "BaseBdev4", 00:09:52.963 "uuid": "dcb1126c-b50c-4cc2-938f-fe09899d3da3", 00:09:52.963 "is_configured": true, 00:09:52.963 "data_offset": 2048, 00:09:52.963 "data_size": 63488 00:09:52.963 } 00:09:52.963 ] 00:09:52.963 }' 00:09:52.963 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:52.963 15:14:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:53.533 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:53.533 15:14:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:53.533 15:14:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:53.533 [2024-11-27 15:14:21.413027] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:53.533 15:14:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:53.533 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:53.533 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:53.533 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:53.533 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:53.533 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:53.533 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:53.533 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:53.533 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:53.533 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:53.533 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:53.533 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:53.533 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:53.533 15:14:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:53.533 15:14:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:53.533 15:14:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:53.533 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:53.533 "name": "Existed_Raid", 00:09:53.533 "uuid": "a9c658f1-072f-4d01-9aea-4f38b2138950", 00:09:53.533 "strip_size_kb": 64, 00:09:53.533 "state": "configuring", 00:09:53.533 "raid_level": "raid0", 00:09:53.533 "superblock": true, 00:09:53.533 "num_base_bdevs": 4, 00:09:53.533 "num_base_bdevs_discovered": 2, 00:09:53.533 "num_base_bdevs_operational": 4, 00:09:53.533 "base_bdevs_list": [ 00:09:53.533 { 00:09:53.533 "name": "BaseBdev1", 00:09:53.533 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:53.533 "is_configured": false, 00:09:53.533 "data_offset": 0, 00:09:53.533 "data_size": 0 00:09:53.533 }, 00:09:53.533 { 00:09:53.533 "name": null, 00:09:53.533 "uuid": "21af45cc-9c06-414c-b0ac-08ef2a4b44c9", 00:09:53.533 "is_configured": false, 00:09:53.533 "data_offset": 0, 00:09:53.533 "data_size": 63488 00:09:53.533 }, 00:09:53.533 { 00:09:53.533 "name": "BaseBdev3", 00:09:53.533 "uuid": "73e8787e-87c5-4d53-ab2d-570781151829", 00:09:53.533 "is_configured": true, 00:09:53.533 "data_offset": 2048, 00:09:53.533 "data_size": 63488 00:09:53.533 }, 00:09:53.533 { 00:09:53.533 "name": "BaseBdev4", 00:09:53.533 "uuid": "dcb1126c-b50c-4cc2-938f-fe09899d3da3", 00:09:53.533 "is_configured": true, 00:09:53.533 "data_offset": 2048, 00:09:53.533 "data_size": 63488 00:09:53.533 } 00:09:53.533 ] 00:09:53.533 }' 00:09:53.533 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:53.533 15:14:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:53.801 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:53.801 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:53.801 15:14:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:53.801 15:14:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:53.801 15:14:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:53.801 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:53.801 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:53.801 15:14:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:53.801 15:14:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:53.801 [2024-11-27 15:14:21.871262] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:53.801 BaseBdev1 00:09:53.801 15:14:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:53.801 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:53.801 15:14:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:53.801 15:14:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:53.801 15:14:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:53.801 15:14:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:53.801 15:14:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:53.801 15:14:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:53.801 15:14:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:53.801 15:14:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:53.802 15:14:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:53.802 15:14:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:53.802 15:14:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:53.802 15:14:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:53.802 [ 00:09:53.802 { 00:09:53.802 "name": "BaseBdev1", 00:09:53.802 "aliases": [ 00:09:53.802 "b32035da-a27f-4fef-9223-6417d0f0d4ff" 00:09:53.802 ], 00:09:53.802 "product_name": "Malloc disk", 00:09:53.802 "block_size": 512, 00:09:53.802 "num_blocks": 65536, 00:09:53.802 "uuid": "b32035da-a27f-4fef-9223-6417d0f0d4ff", 00:09:53.802 "assigned_rate_limits": { 00:09:53.802 "rw_ios_per_sec": 0, 00:09:53.802 "rw_mbytes_per_sec": 0, 00:09:53.802 "r_mbytes_per_sec": 0, 00:09:53.802 "w_mbytes_per_sec": 0 00:09:53.802 }, 00:09:53.802 "claimed": true, 00:09:53.802 "claim_type": "exclusive_write", 00:09:53.802 "zoned": false, 00:09:53.802 "supported_io_types": { 00:09:53.802 "read": true, 00:09:53.802 "write": true, 00:09:53.802 "unmap": true, 00:09:53.802 "flush": true, 00:09:53.802 "reset": true, 00:09:53.802 "nvme_admin": false, 00:09:53.802 "nvme_io": false, 00:09:53.802 "nvme_io_md": false, 00:09:53.802 "write_zeroes": true, 00:09:53.802 "zcopy": true, 00:09:53.802 "get_zone_info": false, 00:09:53.802 "zone_management": false, 00:09:53.802 "zone_append": false, 00:09:53.802 "compare": false, 00:09:54.068 "compare_and_write": false, 00:09:54.068 "abort": true, 00:09:54.068 "seek_hole": false, 00:09:54.068 "seek_data": false, 00:09:54.068 "copy": true, 00:09:54.068 "nvme_iov_md": false 00:09:54.068 }, 00:09:54.068 "memory_domains": [ 00:09:54.068 { 00:09:54.068 "dma_device_id": "system", 00:09:54.068 "dma_device_type": 1 00:09:54.068 }, 00:09:54.068 { 00:09:54.068 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:54.068 "dma_device_type": 2 00:09:54.068 } 00:09:54.068 ], 00:09:54.068 "driver_specific": {} 00:09:54.068 } 00:09:54.068 ] 00:09:54.068 15:14:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.068 15:14:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:54.069 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:54.069 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:54.069 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:54.069 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:54.069 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:54.069 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:54.069 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:54.069 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:54.069 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:54.069 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:54.069 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:54.069 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:54.069 15:14:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.069 15:14:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.069 15:14:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.069 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:54.069 "name": "Existed_Raid", 00:09:54.069 "uuid": "a9c658f1-072f-4d01-9aea-4f38b2138950", 00:09:54.069 "strip_size_kb": 64, 00:09:54.069 "state": "configuring", 00:09:54.069 "raid_level": "raid0", 00:09:54.069 "superblock": true, 00:09:54.069 "num_base_bdevs": 4, 00:09:54.069 "num_base_bdevs_discovered": 3, 00:09:54.069 "num_base_bdevs_operational": 4, 00:09:54.069 "base_bdevs_list": [ 00:09:54.069 { 00:09:54.069 "name": "BaseBdev1", 00:09:54.069 "uuid": "b32035da-a27f-4fef-9223-6417d0f0d4ff", 00:09:54.069 "is_configured": true, 00:09:54.069 "data_offset": 2048, 00:09:54.069 "data_size": 63488 00:09:54.069 }, 00:09:54.069 { 00:09:54.069 "name": null, 00:09:54.069 "uuid": "21af45cc-9c06-414c-b0ac-08ef2a4b44c9", 00:09:54.069 "is_configured": false, 00:09:54.069 "data_offset": 0, 00:09:54.069 "data_size": 63488 00:09:54.069 }, 00:09:54.069 { 00:09:54.069 "name": "BaseBdev3", 00:09:54.069 "uuid": "73e8787e-87c5-4d53-ab2d-570781151829", 00:09:54.069 "is_configured": true, 00:09:54.069 "data_offset": 2048, 00:09:54.069 "data_size": 63488 00:09:54.069 }, 00:09:54.069 { 00:09:54.069 "name": "BaseBdev4", 00:09:54.069 "uuid": "dcb1126c-b50c-4cc2-938f-fe09899d3da3", 00:09:54.069 "is_configured": true, 00:09:54.069 "data_offset": 2048, 00:09:54.069 "data_size": 63488 00:09:54.069 } 00:09:54.069 ] 00:09:54.069 }' 00:09:54.069 15:14:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:54.069 15:14:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.328 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:54.328 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:54.328 15:14:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.328 15:14:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.328 15:14:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.328 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:54.328 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:54.328 15:14:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.328 15:14:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.328 [2024-11-27 15:14:22.402407] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:54.328 15:14:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.328 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:54.328 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:54.328 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:54.328 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:54.328 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:54.328 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:54.328 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:54.328 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:54.328 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:54.328 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:54.328 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:54.328 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:54.328 15:14:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.328 15:14:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.328 15:14:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.588 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:54.588 "name": "Existed_Raid", 00:09:54.588 "uuid": "a9c658f1-072f-4d01-9aea-4f38b2138950", 00:09:54.588 "strip_size_kb": 64, 00:09:54.588 "state": "configuring", 00:09:54.588 "raid_level": "raid0", 00:09:54.588 "superblock": true, 00:09:54.588 "num_base_bdevs": 4, 00:09:54.588 "num_base_bdevs_discovered": 2, 00:09:54.588 "num_base_bdevs_operational": 4, 00:09:54.588 "base_bdevs_list": [ 00:09:54.588 { 00:09:54.588 "name": "BaseBdev1", 00:09:54.588 "uuid": "b32035da-a27f-4fef-9223-6417d0f0d4ff", 00:09:54.588 "is_configured": true, 00:09:54.588 "data_offset": 2048, 00:09:54.588 "data_size": 63488 00:09:54.588 }, 00:09:54.588 { 00:09:54.588 "name": null, 00:09:54.588 "uuid": "21af45cc-9c06-414c-b0ac-08ef2a4b44c9", 00:09:54.588 "is_configured": false, 00:09:54.588 "data_offset": 0, 00:09:54.588 "data_size": 63488 00:09:54.588 }, 00:09:54.588 { 00:09:54.588 "name": null, 00:09:54.588 "uuid": "73e8787e-87c5-4d53-ab2d-570781151829", 00:09:54.588 "is_configured": false, 00:09:54.588 "data_offset": 0, 00:09:54.588 "data_size": 63488 00:09:54.588 }, 00:09:54.588 { 00:09:54.588 "name": "BaseBdev4", 00:09:54.588 "uuid": "dcb1126c-b50c-4cc2-938f-fe09899d3da3", 00:09:54.588 "is_configured": true, 00:09:54.588 "data_offset": 2048, 00:09:54.588 "data_size": 63488 00:09:54.588 } 00:09:54.588 ] 00:09:54.588 }' 00:09:54.588 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:54.588 15:14:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.848 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:54.848 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:54.848 15:14:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.848 15:14:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.848 15:14:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.848 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:54.848 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:54.848 15:14:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.848 15:14:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.848 [2024-11-27 15:14:22.889679] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:54.848 15:14:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.848 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:54.848 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:54.848 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:54.849 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:54.849 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:54.849 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:54.849 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:54.849 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:54.849 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:54.849 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:54.849 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:54.849 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:54.849 15:14:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.849 15:14:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.849 15:14:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.849 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:54.849 "name": "Existed_Raid", 00:09:54.849 "uuid": "a9c658f1-072f-4d01-9aea-4f38b2138950", 00:09:54.849 "strip_size_kb": 64, 00:09:54.849 "state": "configuring", 00:09:54.849 "raid_level": "raid0", 00:09:54.849 "superblock": true, 00:09:54.849 "num_base_bdevs": 4, 00:09:54.849 "num_base_bdevs_discovered": 3, 00:09:54.849 "num_base_bdevs_operational": 4, 00:09:54.849 "base_bdevs_list": [ 00:09:54.849 { 00:09:54.849 "name": "BaseBdev1", 00:09:54.849 "uuid": "b32035da-a27f-4fef-9223-6417d0f0d4ff", 00:09:54.849 "is_configured": true, 00:09:54.849 "data_offset": 2048, 00:09:54.849 "data_size": 63488 00:09:54.849 }, 00:09:54.849 { 00:09:54.849 "name": null, 00:09:54.849 "uuid": "21af45cc-9c06-414c-b0ac-08ef2a4b44c9", 00:09:54.849 "is_configured": false, 00:09:54.849 "data_offset": 0, 00:09:54.849 "data_size": 63488 00:09:54.849 }, 00:09:54.849 { 00:09:54.849 "name": "BaseBdev3", 00:09:54.849 "uuid": "73e8787e-87c5-4d53-ab2d-570781151829", 00:09:54.849 "is_configured": true, 00:09:54.849 "data_offset": 2048, 00:09:54.849 "data_size": 63488 00:09:54.849 }, 00:09:54.849 { 00:09:54.849 "name": "BaseBdev4", 00:09:54.849 "uuid": "dcb1126c-b50c-4cc2-938f-fe09899d3da3", 00:09:54.849 "is_configured": true, 00:09:54.849 "data_offset": 2048, 00:09:54.849 "data_size": 63488 00:09:54.849 } 00:09:54.849 ] 00:09:54.849 }' 00:09:54.849 15:14:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:54.849 15:14:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:55.423 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:55.423 15:14:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.423 15:14:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:55.423 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:55.423 15:14:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.423 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:55.423 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:55.423 15:14:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.423 15:14:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:55.423 [2024-11-27 15:14:23.372891] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:55.423 15:14:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.423 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:55.423 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:55.423 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:55.423 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:55.423 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:55.423 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:55.423 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:55.423 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:55.423 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:55.423 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:55.423 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:55.423 15:14:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.423 15:14:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:55.423 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:55.423 15:14:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.423 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:55.423 "name": "Existed_Raid", 00:09:55.423 "uuid": "a9c658f1-072f-4d01-9aea-4f38b2138950", 00:09:55.423 "strip_size_kb": 64, 00:09:55.423 "state": "configuring", 00:09:55.423 "raid_level": "raid0", 00:09:55.423 "superblock": true, 00:09:55.423 "num_base_bdevs": 4, 00:09:55.423 "num_base_bdevs_discovered": 2, 00:09:55.423 "num_base_bdevs_operational": 4, 00:09:55.423 "base_bdevs_list": [ 00:09:55.423 { 00:09:55.423 "name": null, 00:09:55.423 "uuid": "b32035da-a27f-4fef-9223-6417d0f0d4ff", 00:09:55.423 "is_configured": false, 00:09:55.423 "data_offset": 0, 00:09:55.423 "data_size": 63488 00:09:55.423 }, 00:09:55.423 { 00:09:55.423 "name": null, 00:09:55.423 "uuid": "21af45cc-9c06-414c-b0ac-08ef2a4b44c9", 00:09:55.423 "is_configured": false, 00:09:55.423 "data_offset": 0, 00:09:55.423 "data_size": 63488 00:09:55.423 }, 00:09:55.423 { 00:09:55.423 "name": "BaseBdev3", 00:09:55.423 "uuid": "73e8787e-87c5-4d53-ab2d-570781151829", 00:09:55.423 "is_configured": true, 00:09:55.423 "data_offset": 2048, 00:09:55.423 "data_size": 63488 00:09:55.423 }, 00:09:55.423 { 00:09:55.423 "name": "BaseBdev4", 00:09:55.423 "uuid": "dcb1126c-b50c-4cc2-938f-fe09899d3da3", 00:09:55.423 "is_configured": true, 00:09:55.423 "data_offset": 2048, 00:09:55.423 "data_size": 63488 00:09:55.424 } 00:09:55.424 ] 00:09:55.424 }' 00:09:55.424 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:55.424 15:14:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:55.993 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:55.993 15:14:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.993 15:14:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:55.993 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:55.993 15:14:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.993 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:55.993 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:55.993 15:14:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.993 15:14:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:55.993 [2024-11-27 15:14:23.906588] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:55.993 15:14:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.993 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:55.993 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:55.993 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:55.993 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:55.993 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:55.993 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:55.993 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:55.993 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:55.993 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:55.993 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:55.993 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:55.993 15:14:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.993 15:14:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:55.993 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:55.993 15:14:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.993 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:55.993 "name": "Existed_Raid", 00:09:55.993 "uuid": "a9c658f1-072f-4d01-9aea-4f38b2138950", 00:09:55.993 "strip_size_kb": 64, 00:09:55.993 "state": "configuring", 00:09:55.993 "raid_level": "raid0", 00:09:55.993 "superblock": true, 00:09:55.993 "num_base_bdevs": 4, 00:09:55.993 "num_base_bdevs_discovered": 3, 00:09:55.993 "num_base_bdevs_operational": 4, 00:09:55.993 "base_bdevs_list": [ 00:09:55.993 { 00:09:55.993 "name": null, 00:09:55.993 "uuid": "b32035da-a27f-4fef-9223-6417d0f0d4ff", 00:09:55.993 "is_configured": false, 00:09:55.993 "data_offset": 0, 00:09:55.993 "data_size": 63488 00:09:55.993 }, 00:09:55.993 { 00:09:55.993 "name": "BaseBdev2", 00:09:55.993 "uuid": "21af45cc-9c06-414c-b0ac-08ef2a4b44c9", 00:09:55.993 "is_configured": true, 00:09:55.993 "data_offset": 2048, 00:09:55.993 "data_size": 63488 00:09:55.993 }, 00:09:55.993 { 00:09:55.993 "name": "BaseBdev3", 00:09:55.993 "uuid": "73e8787e-87c5-4d53-ab2d-570781151829", 00:09:55.993 "is_configured": true, 00:09:55.993 "data_offset": 2048, 00:09:55.993 "data_size": 63488 00:09:55.993 }, 00:09:55.993 { 00:09:55.993 "name": "BaseBdev4", 00:09:55.993 "uuid": "dcb1126c-b50c-4cc2-938f-fe09899d3da3", 00:09:55.993 "is_configured": true, 00:09:55.993 "data_offset": 2048, 00:09:55.993 "data_size": 63488 00:09:55.993 } 00:09:55.993 ] 00:09:55.993 }' 00:09:55.993 15:14:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:55.993 15:14:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:56.253 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:56.253 15:14:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.253 15:14:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:56.253 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:56.253 15:14:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.253 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:56.253 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:56.253 15:14:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.253 15:14:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:56.253 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:56.253 15:14:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u b32035da-a27f-4fef-9223-6417d0f0d4ff 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:56.513 [2024-11-27 15:14:24.388850] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:56.513 [2024-11-27 15:14:24.389101] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:09:56.513 NewBaseBdev 00:09:56.513 [2024-11-27 15:14:24.389139] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:56.513 [2024-11-27 15:14:24.389426] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:09:56.513 [2024-11-27 15:14:24.389534] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:09:56.513 [2024-11-27 15:14:24.389545] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006d00 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.513 [2024-11-27 15:14:24.389639] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:56.513 [ 00:09:56.513 { 00:09:56.513 "name": "NewBaseBdev", 00:09:56.513 "aliases": [ 00:09:56.513 "b32035da-a27f-4fef-9223-6417d0f0d4ff" 00:09:56.513 ], 00:09:56.513 "product_name": "Malloc disk", 00:09:56.513 "block_size": 512, 00:09:56.513 "num_blocks": 65536, 00:09:56.513 "uuid": "b32035da-a27f-4fef-9223-6417d0f0d4ff", 00:09:56.513 "assigned_rate_limits": { 00:09:56.513 "rw_ios_per_sec": 0, 00:09:56.513 "rw_mbytes_per_sec": 0, 00:09:56.513 "r_mbytes_per_sec": 0, 00:09:56.513 "w_mbytes_per_sec": 0 00:09:56.513 }, 00:09:56.513 "claimed": true, 00:09:56.513 "claim_type": "exclusive_write", 00:09:56.513 "zoned": false, 00:09:56.513 "supported_io_types": { 00:09:56.513 "read": true, 00:09:56.513 "write": true, 00:09:56.513 "unmap": true, 00:09:56.513 "flush": true, 00:09:56.513 "reset": true, 00:09:56.513 "nvme_admin": false, 00:09:56.513 "nvme_io": false, 00:09:56.513 "nvme_io_md": false, 00:09:56.513 "write_zeroes": true, 00:09:56.513 "zcopy": true, 00:09:56.513 "get_zone_info": false, 00:09:56.513 "zone_management": false, 00:09:56.513 "zone_append": false, 00:09:56.513 "compare": false, 00:09:56.513 "compare_and_write": false, 00:09:56.513 "abort": true, 00:09:56.513 "seek_hole": false, 00:09:56.513 "seek_data": false, 00:09:56.513 "copy": true, 00:09:56.513 "nvme_iov_md": false 00:09:56.513 }, 00:09:56.513 "memory_domains": [ 00:09:56.513 { 00:09:56.513 "dma_device_id": "system", 00:09:56.513 "dma_device_type": 1 00:09:56.513 }, 00:09:56.513 { 00:09:56.513 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:56.513 "dma_device_type": 2 00:09:56.513 } 00:09:56.513 ], 00:09:56.513 "driver_specific": {} 00:09:56.513 } 00:09:56.513 ] 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.513 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:56.513 "name": "Existed_Raid", 00:09:56.513 "uuid": "a9c658f1-072f-4d01-9aea-4f38b2138950", 00:09:56.513 "strip_size_kb": 64, 00:09:56.513 "state": "online", 00:09:56.513 "raid_level": "raid0", 00:09:56.513 "superblock": true, 00:09:56.513 "num_base_bdevs": 4, 00:09:56.514 "num_base_bdevs_discovered": 4, 00:09:56.514 "num_base_bdevs_operational": 4, 00:09:56.514 "base_bdevs_list": [ 00:09:56.514 { 00:09:56.514 "name": "NewBaseBdev", 00:09:56.514 "uuid": "b32035da-a27f-4fef-9223-6417d0f0d4ff", 00:09:56.514 "is_configured": true, 00:09:56.514 "data_offset": 2048, 00:09:56.514 "data_size": 63488 00:09:56.514 }, 00:09:56.514 { 00:09:56.514 "name": "BaseBdev2", 00:09:56.514 "uuid": "21af45cc-9c06-414c-b0ac-08ef2a4b44c9", 00:09:56.514 "is_configured": true, 00:09:56.514 "data_offset": 2048, 00:09:56.514 "data_size": 63488 00:09:56.514 }, 00:09:56.514 { 00:09:56.514 "name": "BaseBdev3", 00:09:56.514 "uuid": "73e8787e-87c5-4d53-ab2d-570781151829", 00:09:56.514 "is_configured": true, 00:09:56.514 "data_offset": 2048, 00:09:56.514 "data_size": 63488 00:09:56.514 }, 00:09:56.514 { 00:09:56.514 "name": "BaseBdev4", 00:09:56.514 "uuid": "dcb1126c-b50c-4cc2-938f-fe09899d3da3", 00:09:56.514 "is_configured": true, 00:09:56.514 "data_offset": 2048, 00:09:56.514 "data_size": 63488 00:09:56.514 } 00:09:56.514 ] 00:09:56.514 }' 00:09:56.514 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:56.514 15:14:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:56.772 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:56.772 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:56.772 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:56.772 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:56.772 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:56.772 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:57.031 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:57.031 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:57.031 15:14:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.031 15:14:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.031 [2024-11-27 15:14:24.884387] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:57.031 15:14:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.031 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:57.031 "name": "Existed_Raid", 00:09:57.031 "aliases": [ 00:09:57.031 "a9c658f1-072f-4d01-9aea-4f38b2138950" 00:09:57.031 ], 00:09:57.031 "product_name": "Raid Volume", 00:09:57.031 "block_size": 512, 00:09:57.031 "num_blocks": 253952, 00:09:57.031 "uuid": "a9c658f1-072f-4d01-9aea-4f38b2138950", 00:09:57.031 "assigned_rate_limits": { 00:09:57.031 "rw_ios_per_sec": 0, 00:09:57.031 "rw_mbytes_per_sec": 0, 00:09:57.031 "r_mbytes_per_sec": 0, 00:09:57.031 "w_mbytes_per_sec": 0 00:09:57.031 }, 00:09:57.031 "claimed": false, 00:09:57.031 "zoned": false, 00:09:57.031 "supported_io_types": { 00:09:57.031 "read": true, 00:09:57.031 "write": true, 00:09:57.031 "unmap": true, 00:09:57.031 "flush": true, 00:09:57.031 "reset": true, 00:09:57.031 "nvme_admin": false, 00:09:57.031 "nvme_io": false, 00:09:57.031 "nvme_io_md": false, 00:09:57.031 "write_zeroes": true, 00:09:57.031 "zcopy": false, 00:09:57.031 "get_zone_info": false, 00:09:57.031 "zone_management": false, 00:09:57.032 "zone_append": false, 00:09:57.032 "compare": false, 00:09:57.032 "compare_and_write": false, 00:09:57.032 "abort": false, 00:09:57.032 "seek_hole": false, 00:09:57.032 "seek_data": false, 00:09:57.032 "copy": false, 00:09:57.032 "nvme_iov_md": false 00:09:57.032 }, 00:09:57.032 "memory_domains": [ 00:09:57.032 { 00:09:57.032 "dma_device_id": "system", 00:09:57.032 "dma_device_type": 1 00:09:57.032 }, 00:09:57.032 { 00:09:57.032 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:57.032 "dma_device_type": 2 00:09:57.032 }, 00:09:57.032 { 00:09:57.032 "dma_device_id": "system", 00:09:57.032 "dma_device_type": 1 00:09:57.032 }, 00:09:57.032 { 00:09:57.032 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:57.032 "dma_device_type": 2 00:09:57.032 }, 00:09:57.032 { 00:09:57.032 "dma_device_id": "system", 00:09:57.032 "dma_device_type": 1 00:09:57.032 }, 00:09:57.032 { 00:09:57.032 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:57.032 "dma_device_type": 2 00:09:57.032 }, 00:09:57.032 { 00:09:57.032 "dma_device_id": "system", 00:09:57.032 "dma_device_type": 1 00:09:57.032 }, 00:09:57.032 { 00:09:57.032 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:57.032 "dma_device_type": 2 00:09:57.032 } 00:09:57.032 ], 00:09:57.032 "driver_specific": { 00:09:57.032 "raid": { 00:09:57.032 "uuid": "a9c658f1-072f-4d01-9aea-4f38b2138950", 00:09:57.032 "strip_size_kb": 64, 00:09:57.032 "state": "online", 00:09:57.032 "raid_level": "raid0", 00:09:57.032 "superblock": true, 00:09:57.032 "num_base_bdevs": 4, 00:09:57.032 "num_base_bdevs_discovered": 4, 00:09:57.032 "num_base_bdevs_operational": 4, 00:09:57.032 "base_bdevs_list": [ 00:09:57.032 { 00:09:57.032 "name": "NewBaseBdev", 00:09:57.032 "uuid": "b32035da-a27f-4fef-9223-6417d0f0d4ff", 00:09:57.032 "is_configured": true, 00:09:57.032 "data_offset": 2048, 00:09:57.032 "data_size": 63488 00:09:57.032 }, 00:09:57.032 { 00:09:57.032 "name": "BaseBdev2", 00:09:57.032 "uuid": "21af45cc-9c06-414c-b0ac-08ef2a4b44c9", 00:09:57.032 "is_configured": true, 00:09:57.032 "data_offset": 2048, 00:09:57.032 "data_size": 63488 00:09:57.032 }, 00:09:57.032 { 00:09:57.032 "name": "BaseBdev3", 00:09:57.032 "uuid": "73e8787e-87c5-4d53-ab2d-570781151829", 00:09:57.032 "is_configured": true, 00:09:57.032 "data_offset": 2048, 00:09:57.032 "data_size": 63488 00:09:57.032 }, 00:09:57.032 { 00:09:57.032 "name": "BaseBdev4", 00:09:57.032 "uuid": "dcb1126c-b50c-4cc2-938f-fe09899d3da3", 00:09:57.032 "is_configured": true, 00:09:57.032 "data_offset": 2048, 00:09:57.032 "data_size": 63488 00:09:57.032 } 00:09:57.032 ] 00:09:57.032 } 00:09:57.032 } 00:09:57.032 }' 00:09:57.032 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:57.032 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:57.032 BaseBdev2 00:09:57.032 BaseBdev3 00:09:57.032 BaseBdev4' 00:09:57.032 15:14:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:57.032 15:14:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:57.032 15:14:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:57.032 15:14:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:57.032 15:14:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.032 15:14:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.032 15:14:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:57.032 15:14:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.032 15:14:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:57.032 15:14:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:57.032 15:14:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:57.032 15:14:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:57.032 15:14:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:57.032 15:14:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.032 15:14:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.032 15:14:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.032 15:14:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:57.032 15:14:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:57.032 15:14:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:57.032 15:14:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:57.032 15:14:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:57.032 15:14:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.032 15:14:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.032 15:14:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.292 15:14:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:57.292 15:14:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:57.292 15:14:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:57.292 15:14:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:57.292 15:14:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:57.292 15:14:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.292 15:14:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.292 15:14:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.292 15:14:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:57.292 15:14:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:57.292 15:14:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:57.292 15:14:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.292 15:14:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.292 [2024-11-27 15:14:25.195546] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:57.292 [2024-11-27 15:14:25.195573] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:57.292 [2024-11-27 15:14:25.195657] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:57.292 [2024-11-27 15:14:25.195723] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:57.292 [2024-11-27 15:14:25.195732] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name Existed_Raid, state offline 00:09:57.292 15:14:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.292 15:14:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 81162 00:09:57.292 15:14:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 81162 ']' 00:09:57.292 15:14:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 81162 00:09:57.292 15:14:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:09:57.292 15:14:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:57.292 15:14:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 81162 00:09:57.292 killing process with pid 81162 00:09:57.292 15:14:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:57.292 15:14:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:57.292 15:14:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 81162' 00:09:57.292 15:14:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 81162 00:09:57.292 [2024-11-27 15:14:25.240577] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:57.292 15:14:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 81162 00:09:57.292 [2024-11-27 15:14:25.281212] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:57.551 15:14:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:09:57.551 00:09:57.551 real 0m9.493s 00:09:57.551 user 0m16.118s 00:09:57.551 sys 0m2.098s 00:09:57.551 ************************************ 00:09:57.551 END TEST raid_state_function_test_sb 00:09:57.551 ************************************ 00:09:57.551 15:14:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:57.551 15:14:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.551 15:14:25 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 4 00:09:57.551 15:14:25 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:09:57.551 15:14:25 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:57.551 15:14:25 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:57.551 ************************************ 00:09:57.551 START TEST raid_superblock_test 00:09:57.551 ************************************ 00:09:57.551 15:14:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid0 4 00:09:57.551 15:14:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:09:57.551 15:14:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:09:57.551 15:14:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:09:57.551 15:14:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:09:57.551 15:14:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:09:57.551 15:14:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:09:57.551 15:14:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:09:57.551 15:14:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:09:57.551 15:14:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:09:57.551 15:14:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:09:57.551 15:14:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:09:57.551 15:14:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:09:57.551 15:14:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:09:57.551 15:14:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:09:57.551 15:14:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:09:57.551 15:14:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:09:57.551 15:14:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=81816 00:09:57.551 15:14:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:09:57.551 15:14:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 81816 00:09:57.551 15:14:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 81816 ']' 00:09:57.551 15:14:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:57.551 15:14:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:57.551 15:14:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:57.551 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:57.551 15:14:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:57.552 15:14:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.810 [2024-11-27 15:14:25.668559] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:09:57.811 [2024-11-27 15:14:25.668788] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid81816 ] 00:09:57.811 [2024-11-27 15:14:25.843086] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:57.811 [2024-11-27 15:14:25.868841] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:57.811 [2024-11-27 15:14:25.912381] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:57.811 [2024-11-27 15:14:25.912475] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.749 malloc1 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.749 [2024-11-27 15:14:26.529018] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:58.749 [2024-11-27 15:14:26.529118] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:58.749 [2024-11-27 15:14:26.529155] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:09:58.749 [2024-11-27 15:14:26.529181] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:58.749 [2024-11-27 15:14:26.531255] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:58.749 [2024-11-27 15:14:26.531295] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:58.749 pt1 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.749 malloc2 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.749 [2024-11-27 15:14:26.557741] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:58.749 [2024-11-27 15:14:26.557842] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:58.749 [2024-11-27 15:14:26.557894] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:09:58.749 [2024-11-27 15:14:26.557938] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:58.749 [2024-11-27 15:14:26.560124] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:58.749 [2024-11-27 15:14:26.560197] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:58.749 pt2 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.749 malloc3 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.749 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.749 [2024-11-27 15:14:26.586425] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:58.749 [2024-11-27 15:14:26.586512] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:58.749 [2024-11-27 15:14:26.586547] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:09:58.749 [2024-11-27 15:14:26.586577] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:58.749 [2024-11-27 15:14:26.588708] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:58.749 [2024-11-27 15:14:26.588795] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:58.749 pt3 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.750 malloc4 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.750 [2024-11-27 15:14:26.623380] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:09:58.750 [2024-11-27 15:14:26.623499] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:58.750 [2024-11-27 15:14:26.623538] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:09:58.750 [2024-11-27 15:14:26.623572] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:58.750 [2024-11-27 15:14:26.625646] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:58.750 [2024-11-27 15:14:26.625716] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:09:58.750 pt4 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.750 [2024-11-27 15:14:26.635439] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:58.750 [2024-11-27 15:14:26.637269] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:58.750 [2024-11-27 15:14:26.637387] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:58.750 [2024-11-27 15:14:26.637458] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:09:58.750 [2024-11-27 15:14:26.637636] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:09:58.750 [2024-11-27 15:14:26.637683] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:58.750 [2024-11-27 15:14:26.637947] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:09:58.750 [2024-11-27 15:14:26.638132] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:09:58.750 [2024-11-27 15:14:26.638173] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:09:58.750 [2024-11-27 15:14:26.638338] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:58.750 "name": "raid_bdev1", 00:09:58.750 "uuid": "7f5b4901-32ff-4655-9da7-07828a0aeac0", 00:09:58.750 "strip_size_kb": 64, 00:09:58.750 "state": "online", 00:09:58.750 "raid_level": "raid0", 00:09:58.750 "superblock": true, 00:09:58.750 "num_base_bdevs": 4, 00:09:58.750 "num_base_bdevs_discovered": 4, 00:09:58.750 "num_base_bdevs_operational": 4, 00:09:58.750 "base_bdevs_list": [ 00:09:58.750 { 00:09:58.750 "name": "pt1", 00:09:58.750 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:58.750 "is_configured": true, 00:09:58.750 "data_offset": 2048, 00:09:58.750 "data_size": 63488 00:09:58.750 }, 00:09:58.750 { 00:09:58.750 "name": "pt2", 00:09:58.750 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:58.750 "is_configured": true, 00:09:58.750 "data_offset": 2048, 00:09:58.750 "data_size": 63488 00:09:58.750 }, 00:09:58.750 { 00:09:58.750 "name": "pt3", 00:09:58.750 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:58.750 "is_configured": true, 00:09:58.750 "data_offset": 2048, 00:09:58.750 "data_size": 63488 00:09:58.750 }, 00:09:58.750 { 00:09:58.750 "name": "pt4", 00:09:58.750 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:58.750 "is_configured": true, 00:09:58.750 "data_offset": 2048, 00:09:58.750 "data_size": 63488 00:09:58.750 } 00:09:58.750 ] 00:09:58.750 }' 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:58.750 15:14:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.010 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:09:59.010 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:59.010 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:59.010 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:59.010 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:59.010 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:59.010 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:59.010 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:59.010 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.010 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.270 [2024-11-27 15:14:27.118893] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:59.270 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.270 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:59.270 "name": "raid_bdev1", 00:09:59.270 "aliases": [ 00:09:59.270 "7f5b4901-32ff-4655-9da7-07828a0aeac0" 00:09:59.270 ], 00:09:59.270 "product_name": "Raid Volume", 00:09:59.270 "block_size": 512, 00:09:59.270 "num_blocks": 253952, 00:09:59.270 "uuid": "7f5b4901-32ff-4655-9da7-07828a0aeac0", 00:09:59.270 "assigned_rate_limits": { 00:09:59.270 "rw_ios_per_sec": 0, 00:09:59.270 "rw_mbytes_per_sec": 0, 00:09:59.270 "r_mbytes_per_sec": 0, 00:09:59.270 "w_mbytes_per_sec": 0 00:09:59.270 }, 00:09:59.270 "claimed": false, 00:09:59.270 "zoned": false, 00:09:59.270 "supported_io_types": { 00:09:59.270 "read": true, 00:09:59.270 "write": true, 00:09:59.270 "unmap": true, 00:09:59.270 "flush": true, 00:09:59.270 "reset": true, 00:09:59.270 "nvme_admin": false, 00:09:59.270 "nvme_io": false, 00:09:59.270 "nvme_io_md": false, 00:09:59.270 "write_zeroes": true, 00:09:59.270 "zcopy": false, 00:09:59.270 "get_zone_info": false, 00:09:59.270 "zone_management": false, 00:09:59.270 "zone_append": false, 00:09:59.270 "compare": false, 00:09:59.270 "compare_and_write": false, 00:09:59.270 "abort": false, 00:09:59.270 "seek_hole": false, 00:09:59.270 "seek_data": false, 00:09:59.270 "copy": false, 00:09:59.270 "nvme_iov_md": false 00:09:59.270 }, 00:09:59.270 "memory_domains": [ 00:09:59.270 { 00:09:59.270 "dma_device_id": "system", 00:09:59.270 "dma_device_type": 1 00:09:59.270 }, 00:09:59.270 { 00:09:59.270 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:59.270 "dma_device_type": 2 00:09:59.270 }, 00:09:59.270 { 00:09:59.270 "dma_device_id": "system", 00:09:59.270 "dma_device_type": 1 00:09:59.270 }, 00:09:59.270 { 00:09:59.270 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:59.270 "dma_device_type": 2 00:09:59.270 }, 00:09:59.270 { 00:09:59.270 "dma_device_id": "system", 00:09:59.270 "dma_device_type": 1 00:09:59.270 }, 00:09:59.270 { 00:09:59.270 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:59.270 "dma_device_type": 2 00:09:59.270 }, 00:09:59.270 { 00:09:59.270 "dma_device_id": "system", 00:09:59.270 "dma_device_type": 1 00:09:59.270 }, 00:09:59.270 { 00:09:59.270 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:59.270 "dma_device_type": 2 00:09:59.270 } 00:09:59.270 ], 00:09:59.270 "driver_specific": { 00:09:59.270 "raid": { 00:09:59.270 "uuid": "7f5b4901-32ff-4655-9da7-07828a0aeac0", 00:09:59.270 "strip_size_kb": 64, 00:09:59.270 "state": "online", 00:09:59.270 "raid_level": "raid0", 00:09:59.270 "superblock": true, 00:09:59.270 "num_base_bdevs": 4, 00:09:59.270 "num_base_bdevs_discovered": 4, 00:09:59.270 "num_base_bdevs_operational": 4, 00:09:59.270 "base_bdevs_list": [ 00:09:59.270 { 00:09:59.270 "name": "pt1", 00:09:59.270 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:59.270 "is_configured": true, 00:09:59.270 "data_offset": 2048, 00:09:59.271 "data_size": 63488 00:09:59.271 }, 00:09:59.271 { 00:09:59.271 "name": "pt2", 00:09:59.271 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:59.271 "is_configured": true, 00:09:59.271 "data_offset": 2048, 00:09:59.271 "data_size": 63488 00:09:59.271 }, 00:09:59.271 { 00:09:59.271 "name": "pt3", 00:09:59.271 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:59.271 "is_configured": true, 00:09:59.271 "data_offset": 2048, 00:09:59.271 "data_size": 63488 00:09:59.271 }, 00:09:59.271 { 00:09:59.271 "name": "pt4", 00:09:59.271 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:59.271 "is_configured": true, 00:09:59.271 "data_offset": 2048, 00:09:59.271 "data_size": 63488 00:09:59.271 } 00:09:59.271 ] 00:09:59.271 } 00:09:59.271 } 00:09:59.271 }' 00:09:59.271 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:59.271 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:59.271 pt2 00:09:59.271 pt3 00:09:59.271 pt4' 00:09:59.271 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:59.271 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:59.271 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:59.271 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:59.271 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.271 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.271 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:59.271 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.271 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:59.271 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:59.271 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:59.271 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:59.271 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.271 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.271 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:59.271 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.271 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:59.271 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:59.271 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:59.271 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:59.271 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:59.271 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.271 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.271 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:09:59.531 [2024-11-27 15:14:27.442294] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=7f5b4901-32ff-4655-9da7-07828a0aeac0 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 7f5b4901-32ff-4655-9da7-07828a0aeac0 ']' 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.531 [2024-11-27 15:14:27.489924] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:59.531 [2024-11-27 15:14:27.489988] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:59.531 [2024-11-27 15:14:27.490102] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:59.531 [2024-11-27 15:14:27.490205] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:59.531 [2024-11-27 15:14:27.490249] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.531 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.792 [2024-11-27 15:14:27.653662] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:09:59.792 [2024-11-27 15:14:27.655462] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:09:59.792 [2024-11-27 15:14:27.655553] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:09:59.792 [2024-11-27 15:14:27.655639] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:09:59.792 [2024-11-27 15:14:27.655783] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:09:59.792 [2024-11-27 15:14:27.655830] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:09:59.792 [2024-11-27 15:14:27.655850] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:09:59.792 [2024-11-27 15:14:27.655866] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:09:59.792 [2024-11-27 15:14:27.655881] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:59.792 [2024-11-27 15:14:27.655891] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state configuring 00:09:59.792 request: 00:09:59.792 { 00:09:59.792 "name": "raid_bdev1", 00:09:59.792 "raid_level": "raid0", 00:09:59.792 "base_bdevs": [ 00:09:59.792 "malloc1", 00:09:59.792 "malloc2", 00:09:59.792 "malloc3", 00:09:59.792 "malloc4" 00:09:59.792 ], 00:09:59.792 "strip_size_kb": 64, 00:09:59.792 "superblock": false, 00:09:59.792 "method": "bdev_raid_create", 00:09:59.792 "req_id": 1 00:09:59.792 } 00:09:59.792 Got JSON-RPC error response 00:09:59.792 response: 00:09:59.792 { 00:09:59.792 "code": -17, 00:09:59.792 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:09:59.792 } 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.792 [2024-11-27 15:14:27.721488] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:59.792 [2024-11-27 15:14:27.721569] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:59.792 [2024-11-27 15:14:27.721606] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:09:59.792 [2024-11-27 15:14:27.721633] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:59.792 [2024-11-27 15:14:27.723722] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:59.792 [2024-11-27 15:14:27.723791] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:59.792 [2024-11-27 15:14:27.723877] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:09:59.792 [2024-11-27 15:14:27.723943] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:59.792 pt1 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 4 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:59.792 "name": "raid_bdev1", 00:09:59.792 "uuid": "7f5b4901-32ff-4655-9da7-07828a0aeac0", 00:09:59.792 "strip_size_kb": 64, 00:09:59.792 "state": "configuring", 00:09:59.792 "raid_level": "raid0", 00:09:59.792 "superblock": true, 00:09:59.792 "num_base_bdevs": 4, 00:09:59.792 "num_base_bdevs_discovered": 1, 00:09:59.792 "num_base_bdevs_operational": 4, 00:09:59.792 "base_bdevs_list": [ 00:09:59.792 { 00:09:59.792 "name": "pt1", 00:09:59.792 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:59.792 "is_configured": true, 00:09:59.792 "data_offset": 2048, 00:09:59.792 "data_size": 63488 00:09:59.792 }, 00:09:59.792 { 00:09:59.792 "name": null, 00:09:59.792 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:59.792 "is_configured": false, 00:09:59.792 "data_offset": 2048, 00:09:59.792 "data_size": 63488 00:09:59.792 }, 00:09:59.792 { 00:09:59.792 "name": null, 00:09:59.792 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:59.792 "is_configured": false, 00:09:59.792 "data_offset": 2048, 00:09:59.792 "data_size": 63488 00:09:59.792 }, 00:09:59.792 { 00:09:59.792 "name": null, 00:09:59.792 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:59.792 "is_configured": false, 00:09:59.792 "data_offset": 2048, 00:09:59.792 "data_size": 63488 00:09:59.792 } 00:09:59.792 ] 00:09:59.792 }' 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:59.792 15:14:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.367 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:10:00.367 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:00.367 15:14:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.367 15:14:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.367 [2024-11-27 15:14:28.176802] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:00.367 [2024-11-27 15:14:28.176859] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:00.367 [2024-11-27 15:14:28.176880] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:10:00.367 [2024-11-27 15:14:28.176889] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:00.367 [2024-11-27 15:14:28.177271] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:00.367 [2024-11-27 15:14:28.177288] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:00.367 [2024-11-27 15:14:28.177358] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:00.367 [2024-11-27 15:14:28.177377] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:00.367 pt2 00:10:00.367 15:14:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.367 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:10:00.367 15:14:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.367 15:14:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.367 [2024-11-27 15:14:28.188787] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:10:00.367 15:14:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.367 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 4 00:10:00.367 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:00.367 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:00.367 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:00.367 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:00.367 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:00.367 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:00.367 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:00.367 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:00.367 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:00.367 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:00.367 15:14:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.367 15:14:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.367 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:00.367 15:14:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.367 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:00.367 "name": "raid_bdev1", 00:10:00.367 "uuid": "7f5b4901-32ff-4655-9da7-07828a0aeac0", 00:10:00.367 "strip_size_kb": 64, 00:10:00.367 "state": "configuring", 00:10:00.367 "raid_level": "raid0", 00:10:00.367 "superblock": true, 00:10:00.367 "num_base_bdevs": 4, 00:10:00.367 "num_base_bdevs_discovered": 1, 00:10:00.367 "num_base_bdevs_operational": 4, 00:10:00.367 "base_bdevs_list": [ 00:10:00.367 { 00:10:00.368 "name": "pt1", 00:10:00.368 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:00.368 "is_configured": true, 00:10:00.368 "data_offset": 2048, 00:10:00.368 "data_size": 63488 00:10:00.368 }, 00:10:00.368 { 00:10:00.368 "name": null, 00:10:00.368 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:00.368 "is_configured": false, 00:10:00.368 "data_offset": 0, 00:10:00.368 "data_size": 63488 00:10:00.368 }, 00:10:00.368 { 00:10:00.368 "name": null, 00:10:00.368 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:00.368 "is_configured": false, 00:10:00.368 "data_offset": 2048, 00:10:00.368 "data_size": 63488 00:10:00.368 }, 00:10:00.368 { 00:10:00.368 "name": null, 00:10:00.368 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:00.368 "is_configured": false, 00:10:00.368 "data_offset": 2048, 00:10:00.368 "data_size": 63488 00:10:00.368 } 00:10:00.368 ] 00:10:00.368 }' 00:10:00.368 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:00.368 15:14:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.640 [2024-11-27 15:14:28.644049] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:00.640 [2024-11-27 15:14:28.644217] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:00.640 [2024-11-27 15:14:28.644256] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:10:00.640 [2024-11-27 15:14:28.644287] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:00.640 [2024-11-27 15:14:28.644716] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:00.640 [2024-11-27 15:14:28.644780] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:00.640 [2024-11-27 15:14:28.644885] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:00.640 [2024-11-27 15:14:28.644958] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:00.640 pt2 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.640 [2024-11-27 15:14:28.655978] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:00.640 [2024-11-27 15:14:28.656071] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:00.640 [2024-11-27 15:14:28.656112] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:10:00.640 [2024-11-27 15:14:28.656147] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:00.640 [2024-11-27 15:14:28.656503] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:00.640 [2024-11-27 15:14:28.656562] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:00.640 [2024-11-27 15:14:28.656652] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:10:00.640 [2024-11-27 15:14:28.656712] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:00.640 pt3 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.640 [2024-11-27 15:14:28.667938] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:00.640 [2024-11-27 15:14:28.668019] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:00.640 [2024-11-27 15:14:28.668036] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:10:00.640 [2024-11-27 15:14:28.668045] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:00.640 [2024-11-27 15:14:28.668350] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:00.640 [2024-11-27 15:14:28.668369] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:00.640 [2024-11-27 15:14:28.668421] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:10:00.640 [2024-11-27 15:14:28.668441] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:00.640 [2024-11-27 15:14:28.668534] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:10:00.640 [2024-11-27 15:14:28.668546] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:00.640 [2024-11-27 15:14:28.668768] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:10:00.640 [2024-11-27 15:14:28.668895] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:10:00.640 [2024-11-27 15:14:28.668903] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:10:00.640 [2024-11-27 15:14:28.669012] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:00.640 pt4 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:00.640 "name": "raid_bdev1", 00:10:00.640 "uuid": "7f5b4901-32ff-4655-9da7-07828a0aeac0", 00:10:00.640 "strip_size_kb": 64, 00:10:00.640 "state": "online", 00:10:00.640 "raid_level": "raid0", 00:10:00.640 "superblock": true, 00:10:00.640 "num_base_bdevs": 4, 00:10:00.640 "num_base_bdevs_discovered": 4, 00:10:00.640 "num_base_bdevs_operational": 4, 00:10:00.640 "base_bdevs_list": [ 00:10:00.640 { 00:10:00.640 "name": "pt1", 00:10:00.640 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:00.640 "is_configured": true, 00:10:00.640 "data_offset": 2048, 00:10:00.640 "data_size": 63488 00:10:00.640 }, 00:10:00.640 { 00:10:00.640 "name": "pt2", 00:10:00.640 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:00.640 "is_configured": true, 00:10:00.640 "data_offset": 2048, 00:10:00.640 "data_size": 63488 00:10:00.640 }, 00:10:00.640 { 00:10:00.640 "name": "pt3", 00:10:00.640 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:00.640 "is_configured": true, 00:10:00.640 "data_offset": 2048, 00:10:00.640 "data_size": 63488 00:10:00.640 }, 00:10:00.640 { 00:10:00.640 "name": "pt4", 00:10:00.640 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:00.640 "is_configured": true, 00:10:00.640 "data_offset": 2048, 00:10:00.640 "data_size": 63488 00:10:00.640 } 00:10:00.640 ] 00:10:00.640 }' 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:00.640 15:14:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.210 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:10:01.210 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:01.210 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:01.210 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:01.210 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:01.210 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:01.210 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:01.210 15:14:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.210 15:14:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.210 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:01.210 [2024-11-27 15:14:29.123550] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:01.210 15:14:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.210 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:01.210 "name": "raid_bdev1", 00:10:01.210 "aliases": [ 00:10:01.210 "7f5b4901-32ff-4655-9da7-07828a0aeac0" 00:10:01.210 ], 00:10:01.210 "product_name": "Raid Volume", 00:10:01.210 "block_size": 512, 00:10:01.210 "num_blocks": 253952, 00:10:01.210 "uuid": "7f5b4901-32ff-4655-9da7-07828a0aeac0", 00:10:01.210 "assigned_rate_limits": { 00:10:01.210 "rw_ios_per_sec": 0, 00:10:01.210 "rw_mbytes_per_sec": 0, 00:10:01.210 "r_mbytes_per_sec": 0, 00:10:01.210 "w_mbytes_per_sec": 0 00:10:01.210 }, 00:10:01.210 "claimed": false, 00:10:01.210 "zoned": false, 00:10:01.210 "supported_io_types": { 00:10:01.210 "read": true, 00:10:01.210 "write": true, 00:10:01.210 "unmap": true, 00:10:01.210 "flush": true, 00:10:01.210 "reset": true, 00:10:01.210 "nvme_admin": false, 00:10:01.210 "nvme_io": false, 00:10:01.210 "nvme_io_md": false, 00:10:01.210 "write_zeroes": true, 00:10:01.210 "zcopy": false, 00:10:01.210 "get_zone_info": false, 00:10:01.210 "zone_management": false, 00:10:01.210 "zone_append": false, 00:10:01.210 "compare": false, 00:10:01.210 "compare_and_write": false, 00:10:01.210 "abort": false, 00:10:01.210 "seek_hole": false, 00:10:01.210 "seek_data": false, 00:10:01.210 "copy": false, 00:10:01.210 "nvme_iov_md": false 00:10:01.210 }, 00:10:01.210 "memory_domains": [ 00:10:01.210 { 00:10:01.210 "dma_device_id": "system", 00:10:01.210 "dma_device_type": 1 00:10:01.210 }, 00:10:01.210 { 00:10:01.210 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:01.210 "dma_device_type": 2 00:10:01.210 }, 00:10:01.210 { 00:10:01.210 "dma_device_id": "system", 00:10:01.210 "dma_device_type": 1 00:10:01.210 }, 00:10:01.210 { 00:10:01.210 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:01.210 "dma_device_type": 2 00:10:01.210 }, 00:10:01.210 { 00:10:01.210 "dma_device_id": "system", 00:10:01.210 "dma_device_type": 1 00:10:01.210 }, 00:10:01.210 { 00:10:01.210 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:01.210 "dma_device_type": 2 00:10:01.210 }, 00:10:01.210 { 00:10:01.210 "dma_device_id": "system", 00:10:01.210 "dma_device_type": 1 00:10:01.210 }, 00:10:01.210 { 00:10:01.210 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:01.210 "dma_device_type": 2 00:10:01.210 } 00:10:01.210 ], 00:10:01.210 "driver_specific": { 00:10:01.210 "raid": { 00:10:01.210 "uuid": "7f5b4901-32ff-4655-9da7-07828a0aeac0", 00:10:01.210 "strip_size_kb": 64, 00:10:01.210 "state": "online", 00:10:01.210 "raid_level": "raid0", 00:10:01.210 "superblock": true, 00:10:01.210 "num_base_bdevs": 4, 00:10:01.210 "num_base_bdevs_discovered": 4, 00:10:01.210 "num_base_bdevs_operational": 4, 00:10:01.210 "base_bdevs_list": [ 00:10:01.210 { 00:10:01.210 "name": "pt1", 00:10:01.210 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:01.210 "is_configured": true, 00:10:01.211 "data_offset": 2048, 00:10:01.211 "data_size": 63488 00:10:01.211 }, 00:10:01.211 { 00:10:01.211 "name": "pt2", 00:10:01.211 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:01.211 "is_configured": true, 00:10:01.211 "data_offset": 2048, 00:10:01.211 "data_size": 63488 00:10:01.211 }, 00:10:01.211 { 00:10:01.211 "name": "pt3", 00:10:01.211 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:01.211 "is_configured": true, 00:10:01.211 "data_offset": 2048, 00:10:01.211 "data_size": 63488 00:10:01.211 }, 00:10:01.211 { 00:10:01.211 "name": "pt4", 00:10:01.211 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:01.211 "is_configured": true, 00:10:01.211 "data_offset": 2048, 00:10:01.211 "data_size": 63488 00:10:01.211 } 00:10:01.211 ] 00:10:01.211 } 00:10:01.211 } 00:10:01.211 }' 00:10:01.211 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:01.211 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:01.211 pt2 00:10:01.211 pt3 00:10:01.211 pt4' 00:10:01.211 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:01.211 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:01.211 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:01.211 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:01.211 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:01.211 15:14:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.211 15:14:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.211 15:14:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.211 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:01.211 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:01.211 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:01.211 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:01.211 15:14:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.211 15:14:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.211 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:01.211 15:14:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:10:01.470 [2024-11-27 15:14:29.426972] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 7f5b4901-32ff-4655-9da7-07828a0aeac0 '!=' 7f5b4901-32ff-4655-9da7-07828a0aeac0 ']' 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 81816 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 81816 ']' 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 81816 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 81816 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 81816' 00:10:01.470 killing process with pid 81816 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 81816 00:10:01.470 [2024-11-27 15:14:29.498301] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:01.470 [2024-11-27 15:14:29.498436] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:01.470 15:14:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 81816 00:10:01.470 [2024-11-27 15:14:29.498535] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:01.470 [2024-11-27 15:14:29.498560] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:10:01.470 [2024-11-27 15:14:29.543338] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:01.729 15:14:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:10:01.729 00:10:01.729 real 0m4.188s 00:10:01.729 user 0m6.595s 00:10:01.729 sys 0m0.949s 00:10:01.729 15:14:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:01.729 15:14:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.729 ************************************ 00:10:01.729 END TEST raid_superblock_test 00:10:01.729 ************************************ 00:10:01.729 15:14:29 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 4 read 00:10:01.729 15:14:29 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:01.729 15:14:29 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:01.729 15:14:29 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:01.729 ************************************ 00:10:01.729 START TEST raid_read_error_test 00:10:01.729 ************************************ 00:10:01.729 15:14:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 4 read 00:10:01.729 15:14:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:10:01.729 15:14:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:10:01.729 15:14:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:10:01.729 15:14:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:01.729 15:14:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:01.729 15:14:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:01.729 15:14:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:01.729 15:14:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:01.729 15:14:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:01.729 15:14:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:01.988 15:14:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:01.988 15:14:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:01.988 15:14:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:01.988 15:14:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:01.988 15:14:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:10:01.988 15:14:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:01.988 15:14:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:01.988 15:14:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:01.988 15:14:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:01.988 15:14:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:01.988 15:14:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:01.988 15:14:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:01.988 15:14:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:01.988 15:14:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:01.988 15:14:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:10:01.988 15:14:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:10:01.988 15:14:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:10:01.988 15:14:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:01.988 15:14:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.JANlWRFvba 00:10:01.988 15:14:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=82064 00:10:01.988 15:14:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 82064 00:10:01.988 15:14:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 82064 ']' 00:10:01.988 15:14:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:01.988 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:01.988 15:14:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:01.988 15:14:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:01.988 15:14:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:01.988 15:14:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.988 15:14:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:01.988 [2024-11-27 15:14:29.929892] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:10:01.988 [2024-11-27 15:14:29.930017] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid82064 ] 00:10:01.988 [2024-11-27 15:14:30.083789] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:02.247 [2024-11-27 15:14:30.110146] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:02.247 [2024-11-27 15:14:30.153014] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:02.247 [2024-11-27 15:14:30.153132] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.815 BaseBdev1_malloc 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.815 true 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.815 [2024-11-27 15:14:30.781281] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:02.815 [2024-11-27 15:14:30.781391] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:02.815 [2024-11-27 15:14:30.781436] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:10:02.815 [2024-11-27 15:14:30.781471] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:02.815 [2024-11-27 15:14:30.783879] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:02.815 [2024-11-27 15:14:30.783964] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:02.815 BaseBdev1 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.815 BaseBdev2_malloc 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.815 true 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.815 [2024-11-27 15:14:30.821938] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:02.815 [2024-11-27 15:14:30.822019] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:02.815 [2024-11-27 15:14:30.822054] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:10:02.815 [2024-11-27 15:14:30.822079] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:02.815 [2024-11-27 15:14:30.824087] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:02.815 [2024-11-27 15:14:30.824159] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:02.815 BaseBdev2 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.815 BaseBdev3_malloc 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.815 true 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.815 [2024-11-27 15:14:30.862448] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:02.815 [2024-11-27 15:14:30.862533] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:02.815 [2024-11-27 15:14:30.862585] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:10:02.815 [2024-11-27 15:14:30.862615] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:02.815 [2024-11-27 15:14:30.864735] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:02.815 [2024-11-27 15:14:30.864819] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:02.815 BaseBdev3 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.815 BaseBdev4_malloc 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.815 true 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.815 [2024-11-27 15:14:30.911153] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:10:02.815 [2024-11-27 15:14:30.911240] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:02.815 [2024-11-27 15:14:30.911280] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:10:02.815 [2024-11-27 15:14:30.911308] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:02.815 [2024-11-27 15:14:30.913429] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:02.815 [2024-11-27 15:14:30.913498] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:10:02.815 BaseBdev4 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:10:02.815 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.816 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.075 [2024-11-27 15:14:30.923189] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:03.075 [2024-11-27 15:14:30.925122] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:03.075 [2024-11-27 15:14:30.925247] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:03.075 [2024-11-27 15:14:30.925320] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:03.075 [2024-11-27 15:14:30.925526] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007080 00:10:03.075 [2024-11-27 15:14:30.925572] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:03.075 [2024-11-27 15:14:30.925836] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006700 00:10:03.075 [2024-11-27 15:14:30.926018] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007080 00:10:03.075 [2024-11-27 15:14:30.926064] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007080 00:10:03.075 [2024-11-27 15:14:30.926232] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:03.075 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.075 15:14:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:10:03.075 15:14:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:03.075 15:14:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:03.075 15:14:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:03.075 15:14:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:03.075 15:14:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:03.075 15:14:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:03.075 15:14:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:03.075 15:14:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:03.075 15:14:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:03.075 15:14:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:03.075 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.075 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.075 15:14:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:03.075 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.075 15:14:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:03.075 "name": "raid_bdev1", 00:10:03.075 "uuid": "ea7a64b9-cfbd-41f2-9c80-7794ea7d9f2b", 00:10:03.075 "strip_size_kb": 64, 00:10:03.075 "state": "online", 00:10:03.075 "raid_level": "raid0", 00:10:03.075 "superblock": true, 00:10:03.075 "num_base_bdevs": 4, 00:10:03.075 "num_base_bdevs_discovered": 4, 00:10:03.075 "num_base_bdevs_operational": 4, 00:10:03.075 "base_bdevs_list": [ 00:10:03.075 { 00:10:03.075 "name": "BaseBdev1", 00:10:03.075 "uuid": "c77c23c9-6615-5618-a29e-0bb973c54bb5", 00:10:03.075 "is_configured": true, 00:10:03.075 "data_offset": 2048, 00:10:03.075 "data_size": 63488 00:10:03.075 }, 00:10:03.075 { 00:10:03.075 "name": "BaseBdev2", 00:10:03.075 "uuid": "6546c1dc-528e-550a-9eef-7d017683007f", 00:10:03.075 "is_configured": true, 00:10:03.075 "data_offset": 2048, 00:10:03.075 "data_size": 63488 00:10:03.075 }, 00:10:03.075 { 00:10:03.075 "name": "BaseBdev3", 00:10:03.075 "uuid": "de8b1c4d-79f4-517e-8854-579aaf3ddc97", 00:10:03.075 "is_configured": true, 00:10:03.075 "data_offset": 2048, 00:10:03.075 "data_size": 63488 00:10:03.075 }, 00:10:03.075 { 00:10:03.075 "name": "BaseBdev4", 00:10:03.075 "uuid": "102b84da-e1f2-536a-bbd8-e8e862d2d73b", 00:10:03.075 "is_configured": true, 00:10:03.075 "data_offset": 2048, 00:10:03.075 "data_size": 63488 00:10:03.075 } 00:10:03.075 ] 00:10:03.075 }' 00:10:03.075 15:14:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:03.075 15:14:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.335 15:14:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:03.335 15:14:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:03.335 [2024-11-27 15:14:31.434763] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000068a0 00:10:04.272 15:14:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:10:04.272 15:14:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.272 15:14:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.272 15:14:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.272 15:14:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:04.272 15:14:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:10:04.272 15:14:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:10:04.272 15:14:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:10:04.272 15:14:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:04.272 15:14:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:04.272 15:14:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:04.272 15:14:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:04.272 15:14:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:04.272 15:14:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:04.272 15:14:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:04.272 15:14:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:04.272 15:14:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:04.531 15:14:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:04.531 15:14:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.531 15:14:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:04.531 15:14:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.531 15:14:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.531 15:14:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:04.531 "name": "raid_bdev1", 00:10:04.531 "uuid": "ea7a64b9-cfbd-41f2-9c80-7794ea7d9f2b", 00:10:04.531 "strip_size_kb": 64, 00:10:04.531 "state": "online", 00:10:04.531 "raid_level": "raid0", 00:10:04.531 "superblock": true, 00:10:04.531 "num_base_bdevs": 4, 00:10:04.531 "num_base_bdevs_discovered": 4, 00:10:04.531 "num_base_bdevs_operational": 4, 00:10:04.531 "base_bdevs_list": [ 00:10:04.531 { 00:10:04.531 "name": "BaseBdev1", 00:10:04.531 "uuid": "c77c23c9-6615-5618-a29e-0bb973c54bb5", 00:10:04.531 "is_configured": true, 00:10:04.531 "data_offset": 2048, 00:10:04.531 "data_size": 63488 00:10:04.531 }, 00:10:04.531 { 00:10:04.531 "name": "BaseBdev2", 00:10:04.531 "uuid": "6546c1dc-528e-550a-9eef-7d017683007f", 00:10:04.531 "is_configured": true, 00:10:04.531 "data_offset": 2048, 00:10:04.531 "data_size": 63488 00:10:04.531 }, 00:10:04.531 { 00:10:04.531 "name": "BaseBdev3", 00:10:04.531 "uuid": "de8b1c4d-79f4-517e-8854-579aaf3ddc97", 00:10:04.531 "is_configured": true, 00:10:04.531 "data_offset": 2048, 00:10:04.531 "data_size": 63488 00:10:04.531 }, 00:10:04.531 { 00:10:04.531 "name": "BaseBdev4", 00:10:04.531 "uuid": "102b84da-e1f2-536a-bbd8-e8e862d2d73b", 00:10:04.531 "is_configured": true, 00:10:04.531 "data_offset": 2048, 00:10:04.531 "data_size": 63488 00:10:04.531 } 00:10:04.531 ] 00:10:04.531 }' 00:10:04.531 15:14:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:04.531 15:14:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.791 15:14:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:04.791 15:14:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.791 15:14:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.791 [2024-11-27 15:14:32.770417] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:04.791 [2024-11-27 15:14:32.770495] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:04.791 [2024-11-27 15:14:32.773023] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:04.791 [2024-11-27 15:14:32.773112] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:04.791 [2024-11-27 15:14:32.773173] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:04.791 [2024-11-27 15:14:32.773227] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007080 name raid_bdev1, state offline 00:10:04.791 { 00:10:04.791 "results": [ 00:10:04.791 { 00:10:04.791 "job": "raid_bdev1", 00:10:04.791 "core_mask": "0x1", 00:10:04.791 "workload": "randrw", 00:10:04.791 "percentage": 50, 00:10:04.791 "status": "finished", 00:10:04.791 "queue_depth": 1, 00:10:04.791 "io_size": 131072, 00:10:04.791 "runtime": 1.336388, 00:10:04.791 "iops": 16129.297778788796, 00:10:04.791 "mibps": 2016.1622223485995, 00:10:04.791 "io_failed": 1, 00:10:04.791 "io_timeout": 0, 00:10:04.791 "avg_latency_us": 85.74379412696574, 00:10:04.791 "min_latency_us": 26.382532751091702, 00:10:04.791 "max_latency_us": 1402.2986899563318 00:10:04.791 } 00:10:04.791 ], 00:10:04.791 "core_count": 1 00:10:04.791 } 00:10:04.791 15:14:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.791 15:14:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 82064 00:10:04.791 15:14:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 82064 ']' 00:10:04.791 15:14:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 82064 00:10:04.791 15:14:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:10:04.791 15:14:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:04.791 15:14:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 82064 00:10:04.791 15:14:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:04.791 15:14:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:04.791 15:14:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 82064' 00:10:04.791 killing process with pid 82064 00:10:04.791 15:14:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 82064 00:10:04.791 [2024-11-27 15:14:32.820545] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:04.791 15:14:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 82064 00:10:04.791 [2024-11-27 15:14:32.855826] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:05.051 15:14:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.JANlWRFvba 00:10:05.051 15:14:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:05.051 15:14:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:05.051 15:14:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.75 00:10:05.051 15:14:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:10:05.051 15:14:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:05.051 15:14:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:05.051 15:14:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.75 != \0\.\0\0 ]] 00:10:05.051 00:10:05.051 real 0m3.249s 00:10:05.051 user 0m4.059s 00:10:05.051 sys 0m0.532s 00:10:05.051 15:14:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:05.051 ************************************ 00:10:05.051 END TEST raid_read_error_test 00:10:05.051 ************************************ 00:10:05.051 15:14:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.051 15:14:33 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 4 write 00:10:05.051 15:14:33 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:05.051 15:14:33 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:05.051 15:14:33 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:05.051 ************************************ 00:10:05.051 START TEST raid_write_error_test 00:10:05.051 ************************************ 00:10:05.051 15:14:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 4 write 00:10:05.051 15:14:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:10:05.051 15:14:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:10:05.051 15:14:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:10:05.051 15:14:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:05.051 15:14:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:05.051 15:14:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:05.051 15:14:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:05.051 15:14:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:05.051 15:14:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:05.310 15:14:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:05.310 15:14:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:05.310 15:14:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:05.310 15:14:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:05.310 15:14:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:05.310 15:14:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:10:05.310 15:14:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:05.310 15:14:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:05.310 15:14:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:05.310 15:14:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:05.310 15:14:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:05.310 15:14:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:05.310 15:14:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:05.310 15:14:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:05.310 15:14:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:05.310 15:14:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:10:05.310 15:14:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:10:05.310 15:14:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:10:05.310 15:14:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:05.310 15:14:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.CeAaIpoA0o 00:10:05.310 15:14:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=82194 00:10:05.310 15:14:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 82194 00:10:05.310 15:14:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:05.310 15:14:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 82194 ']' 00:10:05.310 15:14:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:05.310 15:14:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:05.310 15:14:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:05.310 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:05.310 15:14:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:05.310 15:14:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.310 [2024-11-27 15:14:33.262541] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:10:05.310 [2024-11-27 15:14:33.262764] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid82194 ] 00:10:05.569 [2024-11-27 15:14:33.435179] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:05.569 [2024-11-27 15:14:33.461729] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:05.569 [2024-11-27 15:14:33.504787] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:05.569 [2024-11-27 15:14:33.504896] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.137 BaseBdev1_malloc 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.137 true 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.137 [2024-11-27 15:14:34.124812] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:06.137 [2024-11-27 15:14:34.124923] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:06.137 [2024-11-27 15:14:34.124971] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:10:06.137 [2024-11-27 15:14:34.125010] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:06.137 [2024-11-27 15:14:34.127102] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:06.137 [2024-11-27 15:14:34.127173] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:06.137 BaseBdev1 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.137 BaseBdev2_malloc 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.137 true 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.137 [2024-11-27 15:14:34.165444] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:06.137 [2024-11-27 15:14:34.165529] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:06.137 [2024-11-27 15:14:34.165564] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:10:06.137 [2024-11-27 15:14:34.165592] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:06.137 [2024-11-27 15:14:34.167615] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:06.137 [2024-11-27 15:14:34.167684] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:06.137 BaseBdev2 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.137 BaseBdev3_malloc 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.137 true 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.137 [2024-11-27 15:14:34.206021] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:06.137 [2024-11-27 15:14:34.206103] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:06.137 [2024-11-27 15:14:34.206139] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:10:06.137 [2024-11-27 15:14:34.206167] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:06.137 [2024-11-27 15:14:34.208154] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:06.137 [2024-11-27 15:14:34.208231] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:06.137 BaseBdev3 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.137 BaseBdev4_malloc 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.137 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.396 true 00:10:06.396 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.396 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:10:06.396 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.396 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.396 [2024-11-27 15:14:34.256735] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:10:06.396 [2024-11-27 15:14:34.256820] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:06.396 [2024-11-27 15:14:34.256858] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:10:06.396 [2024-11-27 15:14:34.256869] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:06.396 [2024-11-27 15:14:34.258821] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:06.396 [2024-11-27 15:14:34.258858] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:10:06.396 BaseBdev4 00:10:06.396 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.396 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:10:06.396 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.396 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.396 [2024-11-27 15:14:34.268784] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:06.396 [2024-11-27 15:14:34.270598] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:06.396 [2024-11-27 15:14:34.270726] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:06.396 [2024-11-27 15:14:34.270799] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:06.396 [2024-11-27 15:14:34.271036] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007080 00:10:06.396 [2024-11-27 15:14:34.271087] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:06.396 [2024-11-27 15:14:34.271353] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006700 00:10:06.396 [2024-11-27 15:14:34.271508] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007080 00:10:06.396 [2024-11-27 15:14:34.271551] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007080 00:10:06.396 [2024-11-27 15:14:34.271740] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:06.396 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.396 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:10:06.396 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:06.396 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:06.396 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:06.396 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:06.396 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:06.396 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:06.396 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:06.396 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:06.396 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:06.396 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:06.396 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:06.397 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.397 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.397 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.397 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:06.397 "name": "raid_bdev1", 00:10:06.397 "uuid": "3fb564be-94eb-4fa3-9d49-b1af071582de", 00:10:06.397 "strip_size_kb": 64, 00:10:06.397 "state": "online", 00:10:06.397 "raid_level": "raid0", 00:10:06.397 "superblock": true, 00:10:06.397 "num_base_bdevs": 4, 00:10:06.397 "num_base_bdevs_discovered": 4, 00:10:06.397 "num_base_bdevs_operational": 4, 00:10:06.397 "base_bdevs_list": [ 00:10:06.397 { 00:10:06.397 "name": "BaseBdev1", 00:10:06.397 "uuid": "802d9f98-2727-5570-8b9f-f05b92cc444c", 00:10:06.397 "is_configured": true, 00:10:06.397 "data_offset": 2048, 00:10:06.397 "data_size": 63488 00:10:06.397 }, 00:10:06.397 { 00:10:06.397 "name": "BaseBdev2", 00:10:06.397 "uuid": "f81a0a1b-18ca-554c-af8b-aeaeb21f27e4", 00:10:06.397 "is_configured": true, 00:10:06.397 "data_offset": 2048, 00:10:06.397 "data_size": 63488 00:10:06.397 }, 00:10:06.397 { 00:10:06.397 "name": "BaseBdev3", 00:10:06.397 "uuid": "95da8c9a-1d3e-53b4-bb50-2d03e62bd256", 00:10:06.397 "is_configured": true, 00:10:06.397 "data_offset": 2048, 00:10:06.397 "data_size": 63488 00:10:06.397 }, 00:10:06.397 { 00:10:06.397 "name": "BaseBdev4", 00:10:06.397 "uuid": "3a3dc554-2c80-51f4-ad2b-9f683c24c64d", 00:10:06.397 "is_configured": true, 00:10:06.397 "data_offset": 2048, 00:10:06.397 "data_size": 63488 00:10:06.397 } 00:10:06.397 ] 00:10:06.397 }' 00:10:06.397 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:06.397 15:14:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.655 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:06.655 15:14:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:06.914 [2024-11-27 15:14:34.808210] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000068a0 00:10:07.862 15:14:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:10:07.862 15:14:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.862 15:14:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.862 15:14:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.862 15:14:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:07.862 15:14:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:10:07.862 15:14:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:10:07.862 15:14:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:10:07.863 15:14:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:07.863 15:14:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:07.863 15:14:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:07.863 15:14:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:07.863 15:14:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:07.863 15:14:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:07.863 15:14:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:07.863 15:14:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:07.863 15:14:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:07.863 15:14:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:07.863 15:14:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:07.863 15:14:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.863 15:14:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.863 15:14:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.863 15:14:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:07.863 "name": "raid_bdev1", 00:10:07.863 "uuid": "3fb564be-94eb-4fa3-9d49-b1af071582de", 00:10:07.863 "strip_size_kb": 64, 00:10:07.863 "state": "online", 00:10:07.863 "raid_level": "raid0", 00:10:07.863 "superblock": true, 00:10:07.863 "num_base_bdevs": 4, 00:10:07.863 "num_base_bdevs_discovered": 4, 00:10:07.863 "num_base_bdevs_operational": 4, 00:10:07.863 "base_bdevs_list": [ 00:10:07.863 { 00:10:07.863 "name": "BaseBdev1", 00:10:07.863 "uuid": "802d9f98-2727-5570-8b9f-f05b92cc444c", 00:10:07.863 "is_configured": true, 00:10:07.863 "data_offset": 2048, 00:10:07.863 "data_size": 63488 00:10:07.863 }, 00:10:07.863 { 00:10:07.863 "name": "BaseBdev2", 00:10:07.863 "uuid": "f81a0a1b-18ca-554c-af8b-aeaeb21f27e4", 00:10:07.863 "is_configured": true, 00:10:07.863 "data_offset": 2048, 00:10:07.863 "data_size": 63488 00:10:07.863 }, 00:10:07.863 { 00:10:07.863 "name": "BaseBdev3", 00:10:07.863 "uuid": "95da8c9a-1d3e-53b4-bb50-2d03e62bd256", 00:10:07.863 "is_configured": true, 00:10:07.863 "data_offset": 2048, 00:10:07.863 "data_size": 63488 00:10:07.863 }, 00:10:07.863 { 00:10:07.863 "name": "BaseBdev4", 00:10:07.863 "uuid": "3a3dc554-2c80-51f4-ad2b-9f683c24c64d", 00:10:07.863 "is_configured": true, 00:10:07.863 "data_offset": 2048, 00:10:07.863 "data_size": 63488 00:10:07.863 } 00:10:07.863 ] 00:10:07.863 }' 00:10:07.863 15:14:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:07.863 15:14:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:08.122 15:14:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:08.122 15:14:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:08.122 15:14:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:08.122 [2024-11-27 15:14:36.168165] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:08.122 [2024-11-27 15:14:36.168244] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:08.122 [2024-11-27 15:14:36.170841] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:08.122 [2024-11-27 15:14:36.170971] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:08.122 [2024-11-27 15:14:36.171046] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:08.122 [2024-11-27 15:14:36.171095] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007080 name raid_bdev1, state offline 00:10:08.122 { 00:10:08.122 "results": [ 00:10:08.122 { 00:10:08.122 "job": "raid_bdev1", 00:10:08.122 "core_mask": "0x1", 00:10:08.122 "workload": "randrw", 00:10:08.122 "percentage": 50, 00:10:08.122 "status": "finished", 00:10:08.122 "queue_depth": 1, 00:10:08.122 "io_size": 131072, 00:10:08.122 "runtime": 1.360961, 00:10:08.122 "iops": 15934.328757400102, 00:10:08.122 "mibps": 1991.7910946750128, 00:10:08.122 "io_failed": 1, 00:10:08.122 "io_timeout": 0, 00:10:08.122 "avg_latency_us": 86.74370829283556, 00:10:08.122 "min_latency_us": 26.270742358078603, 00:10:08.122 "max_latency_us": 1409.4532751091704 00:10:08.122 } 00:10:08.122 ], 00:10:08.122 "core_count": 1 00:10:08.122 } 00:10:08.122 15:14:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:08.122 15:14:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 82194 00:10:08.122 15:14:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 82194 ']' 00:10:08.122 15:14:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 82194 00:10:08.122 15:14:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:10:08.122 15:14:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:08.122 15:14:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 82194 00:10:08.122 15:14:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:08.122 killing process with pid 82194 00:10:08.122 15:14:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:08.122 15:14:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 82194' 00:10:08.122 15:14:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 82194 00:10:08.122 [2024-11-27 15:14:36.224454] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:08.123 15:14:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 82194 00:10:08.382 [2024-11-27 15:14:36.260543] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:08.382 15:14:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:08.382 15:14:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.CeAaIpoA0o 00:10:08.382 15:14:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:08.382 15:14:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.73 00:10:08.382 15:14:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:10:08.382 15:14:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:08.382 15:14:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:08.382 ************************************ 00:10:08.382 END TEST raid_write_error_test 00:10:08.382 ************************************ 00:10:08.382 15:14:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.73 != \0\.\0\0 ]] 00:10:08.382 00:10:08.382 real 0m3.331s 00:10:08.382 user 0m4.223s 00:10:08.382 sys 0m0.543s 00:10:08.382 15:14:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:08.382 15:14:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:08.641 15:14:36 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:10:08.641 15:14:36 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 4 false 00:10:08.641 15:14:36 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:08.641 15:14:36 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:08.641 15:14:36 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:08.641 ************************************ 00:10:08.641 START TEST raid_state_function_test 00:10:08.641 ************************************ 00:10:08.641 15:14:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 4 false 00:10:08.641 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:10:08.641 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:10:08.641 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:10:08.641 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:10:08.641 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:10:08.641 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:08.641 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:10:08.641 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:08.641 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:08.641 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:10:08.641 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:08.641 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:08.641 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:10:08.641 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:08.641 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:08.641 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:10:08.641 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:08.641 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:08.641 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:08.641 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:10:08.641 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:10:08.641 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:10:08.641 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:10:08.641 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:10:08.641 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:10:08.642 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:10:08.642 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:10:08.642 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:10:08.642 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:10:08.642 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=82321 00:10:08.642 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:10:08.642 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 82321' 00:10:08.642 Process raid pid: 82321 00:10:08.642 15:14:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 82321 00:10:08.642 15:14:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 82321 ']' 00:10:08.642 15:14:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:08.642 15:14:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:08.642 15:14:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:08.642 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:08.642 15:14:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:08.642 15:14:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:08.642 [2024-11-27 15:14:36.650235] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:10:08.642 [2024-11-27 15:14:36.650434] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:08.901 [2024-11-27 15:14:36.822121] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:08.901 [2024-11-27 15:14:36.847501] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:08.901 [2024-11-27 15:14:36.890403] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:08.901 [2024-11-27 15:14:36.890515] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:09.468 15:14:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:09.468 15:14:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:10:09.468 15:14:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:09.468 15:14:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.468 15:14:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.468 [2024-11-27 15:14:37.533380] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:09.468 [2024-11-27 15:14:37.533444] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:09.468 [2024-11-27 15:14:37.533453] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:09.468 [2024-11-27 15:14:37.533463] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:09.468 [2024-11-27 15:14:37.533469] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:09.468 [2024-11-27 15:14:37.533479] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:09.468 [2024-11-27 15:14:37.533485] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:09.468 [2024-11-27 15:14:37.533493] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:09.468 15:14:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.468 15:14:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:09.468 15:14:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:09.468 15:14:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:09.468 15:14:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:09.468 15:14:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:09.468 15:14:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:09.468 15:14:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:09.468 15:14:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:09.468 15:14:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:09.468 15:14:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:09.468 15:14:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:09.468 15:14:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.468 15:14:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.468 15:14:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:09.469 15:14:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.727 15:14:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:09.727 "name": "Existed_Raid", 00:10:09.727 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:09.727 "strip_size_kb": 64, 00:10:09.727 "state": "configuring", 00:10:09.727 "raid_level": "concat", 00:10:09.727 "superblock": false, 00:10:09.727 "num_base_bdevs": 4, 00:10:09.727 "num_base_bdevs_discovered": 0, 00:10:09.727 "num_base_bdevs_operational": 4, 00:10:09.727 "base_bdevs_list": [ 00:10:09.727 { 00:10:09.727 "name": "BaseBdev1", 00:10:09.727 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:09.727 "is_configured": false, 00:10:09.727 "data_offset": 0, 00:10:09.727 "data_size": 0 00:10:09.727 }, 00:10:09.727 { 00:10:09.727 "name": "BaseBdev2", 00:10:09.727 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:09.727 "is_configured": false, 00:10:09.727 "data_offset": 0, 00:10:09.727 "data_size": 0 00:10:09.727 }, 00:10:09.727 { 00:10:09.727 "name": "BaseBdev3", 00:10:09.727 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:09.727 "is_configured": false, 00:10:09.727 "data_offset": 0, 00:10:09.727 "data_size": 0 00:10:09.727 }, 00:10:09.727 { 00:10:09.727 "name": "BaseBdev4", 00:10:09.727 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:09.727 "is_configured": false, 00:10:09.727 "data_offset": 0, 00:10:09.727 "data_size": 0 00:10:09.727 } 00:10:09.727 ] 00:10:09.727 }' 00:10:09.727 15:14:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:09.727 15:14:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.986 15:14:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:09.986 15:14:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.986 15:14:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.986 [2024-11-27 15:14:37.960562] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:09.986 [2024-11-27 15:14:37.960651] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:10:09.986 15:14:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.986 15:14:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:09.986 15:14:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.986 15:14:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.986 [2024-11-27 15:14:37.972565] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:09.986 [2024-11-27 15:14:37.972645] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:09.986 [2024-11-27 15:14:37.972674] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:09.986 [2024-11-27 15:14:37.972697] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:09.986 [2024-11-27 15:14:37.972716] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:09.986 [2024-11-27 15:14:37.972737] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:09.986 [2024-11-27 15:14:37.972755] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:09.986 [2024-11-27 15:14:37.972776] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:09.986 15:14:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.986 15:14:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:09.986 15:14:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.986 15:14:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.986 [2024-11-27 15:14:37.993653] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:09.986 BaseBdev1 00:10:09.986 15:14:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.986 15:14:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:10:09.986 15:14:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:09.986 15:14:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:09.986 15:14:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:09.986 15:14:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:09.986 15:14:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:09.986 15:14:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:09.986 15:14:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.986 15:14:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.986 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.986 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:09.986 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.986 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.986 [ 00:10:09.986 { 00:10:09.986 "name": "BaseBdev1", 00:10:09.986 "aliases": [ 00:10:09.986 "10d0ce0e-51ff-476f-b148-187393f14178" 00:10:09.986 ], 00:10:09.986 "product_name": "Malloc disk", 00:10:09.986 "block_size": 512, 00:10:09.986 "num_blocks": 65536, 00:10:09.986 "uuid": "10d0ce0e-51ff-476f-b148-187393f14178", 00:10:09.986 "assigned_rate_limits": { 00:10:09.986 "rw_ios_per_sec": 0, 00:10:09.986 "rw_mbytes_per_sec": 0, 00:10:09.986 "r_mbytes_per_sec": 0, 00:10:09.986 "w_mbytes_per_sec": 0 00:10:09.986 }, 00:10:09.986 "claimed": true, 00:10:09.986 "claim_type": "exclusive_write", 00:10:09.986 "zoned": false, 00:10:09.986 "supported_io_types": { 00:10:09.986 "read": true, 00:10:09.986 "write": true, 00:10:09.986 "unmap": true, 00:10:09.986 "flush": true, 00:10:09.986 "reset": true, 00:10:09.986 "nvme_admin": false, 00:10:09.986 "nvme_io": false, 00:10:09.986 "nvme_io_md": false, 00:10:09.986 "write_zeroes": true, 00:10:09.986 "zcopy": true, 00:10:09.986 "get_zone_info": false, 00:10:09.986 "zone_management": false, 00:10:09.986 "zone_append": false, 00:10:09.986 "compare": false, 00:10:09.986 "compare_and_write": false, 00:10:09.986 "abort": true, 00:10:09.986 "seek_hole": false, 00:10:09.986 "seek_data": false, 00:10:09.986 "copy": true, 00:10:09.986 "nvme_iov_md": false 00:10:09.986 }, 00:10:09.986 "memory_domains": [ 00:10:09.986 { 00:10:09.986 "dma_device_id": "system", 00:10:09.986 "dma_device_type": 1 00:10:09.986 }, 00:10:09.986 { 00:10:09.986 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:09.986 "dma_device_type": 2 00:10:09.986 } 00:10:09.986 ], 00:10:09.986 "driver_specific": {} 00:10:09.986 } 00:10:09.986 ] 00:10:09.986 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.986 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:09.986 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:09.986 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:09.986 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:09.986 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:09.986 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:09.986 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:09.986 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:09.986 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:09.986 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:09.986 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:09.986 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:09.986 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.986 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.986 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:09.986 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.986 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:09.986 "name": "Existed_Raid", 00:10:09.986 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:09.986 "strip_size_kb": 64, 00:10:09.986 "state": "configuring", 00:10:09.986 "raid_level": "concat", 00:10:09.986 "superblock": false, 00:10:09.986 "num_base_bdevs": 4, 00:10:09.986 "num_base_bdevs_discovered": 1, 00:10:09.986 "num_base_bdevs_operational": 4, 00:10:09.986 "base_bdevs_list": [ 00:10:09.986 { 00:10:09.986 "name": "BaseBdev1", 00:10:09.986 "uuid": "10d0ce0e-51ff-476f-b148-187393f14178", 00:10:09.986 "is_configured": true, 00:10:09.986 "data_offset": 0, 00:10:09.986 "data_size": 65536 00:10:09.986 }, 00:10:09.986 { 00:10:09.986 "name": "BaseBdev2", 00:10:09.986 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:09.986 "is_configured": false, 00:10:09.986 "data_offset": 0, 00:10:09.986 "data_size": 0 00:10:09.986 }, 00:10:09.986 { 00:10:09.986 "name": "BaseBdev3", 00:10:09.986 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:09.986 "is_configured": false, 00:10:09.986 "data_offset": 0, 00:10:09.986 "data_size": 0 00:10:09.986 }, 00:10:09.986 { 00:10:09.986 "name": "BaseBdev4", 00:10:09.986 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:09.986 "is_configured": false, 00:10:09.986 "data_offset": 0, 00:10:09.986 "data_size": 0 00:10:09.986 } 00:10:09.986 ] 00:10:09.986 }' 00:10:09.987 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:09.987 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.554 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:10.554 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.554 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.554 [2024-11-27 15:14:38.448968] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:10.554 [2024-11-27 15:14:38.449010] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:10:10.554 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.554 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:10.554 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.554 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.554 [2024-11-27 15:14:38.460995] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:10.554 [2024-11-27 15:14:38.462784] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:10.554 [2024-11-27 15:14:38.462827] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:10.554 [2024-11-27 15:14:38.462836] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:10.554 [2024-11-27 15:14:38.462844] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:10.554 [2024-11-27 15:14:38.462850] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:10.554 [2024-11-27 15:14:38.462858] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:10.554 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.554 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:10:10.554 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:10.554 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:10.554 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:10.554 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:10.554 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:10.554 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:10.554 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:10.554 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:10.554 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:10.554 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:10.554 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:10.554 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:10.554 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:10.554 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.554 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.554 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.554 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:10.554 "name": "Existed_Raid", 00:10:10.554 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:10.554 "strip_size_kb": 64, 00:10:10.554 "state": "configuring", 00:10:10.554 "raid_level": "concat", 00:10:10.554 "superblock": false, 00:10:10.554 "num_base_bdevs": 4, 00:10:10.554 "num_base_bdevs_discovered": 1, 00:10:10.554 "num_base_bdevs_operational": 4, 00:10:10.554 "base_bdevs_list": [ 00:10:10.554 { 00:10:10.554 "name": "BaseBdev1", 00:10:10.554 "uuid": "10d0ce0e-51ff-476f-b148-187393f14178", 00:10:10.554 "is_configured": true, 00:10:10.554 "data_offset": 0, 00:10:10.554 "data_size": 65536 00:10:10.554 }, 00:10:10.554 { 00:10:10.554 "name": "BaseBdev2", 00:10:10.554 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:10.554 "is_configured": false, 00:10:10.554 "data_offset": 0, 00:10:10.554 "data_size": 0 00:10:10.554 }, 00:10:10.554 { 00:10:10.554 "name": "BaseBdev3", 00:10:10.554 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:10.554 "is_configured": false, 00:10:10.554 "data_offset": 0, 00:10:10.554 "data_size": 0 00:10:10.554 }, 00:10:10.554 { 00:10:10.554 "name": "BaseBdev4", 00:10:10.554 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:10.554 "is_configured": false, 00:10:10.554 "data_offset": 0, 00:10:10.554 "data_size": 0 00:10:10.554 } 00:10:10.554 ] 00:10:10.554 }' 00:10:10.554 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:10.554 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.814 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:10.814 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.814 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.814 [2024-11-27 15:14:38.867320] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:10.814 BaseBdev2 00:10:10.814 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.814 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:10:10.814 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:10.814 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:10.814 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:10.814 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:10.814 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:10.814 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:10.814 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.814 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.814 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.814 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:10.814 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.814 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.814 [ 00:10:10.814 { 00:10:10.814 "name": "BaseBdev2", 00:10:10.814 "aliases": [ 00:10:10.814 "357aadef-565c-49b4-9313-72cf3da4824b" 00:10:10.814 ], 00:10:10.814 "product_name": "Malloc disk", 00:10:10.814 "block_size": 512, 00:10:10.814 "num_blocks": 65536, 00:10:10.814 "uuid": "357aadef-565c-49b4-9313-72cf3da4824b", 00:10:10.814 "assigned_rate_limits": { 00:10:10.814 "rw_ios_per_sec": 0, 00:10:10.814 "rw_mbytes_per_sec": 0, 00:10:10.814 "r_mbytes_per_sec": 0, 00:10:10.814 "w_mbytes_per_sec": 0 00:10:10.814 }, 00:10:10.814 "claimed": true, 00:10:10.814 "claim_type": "exclusive_write", 00:10:10.814 "zoned": false, 00:10:10.814 "supported_io_types": { 00:10:10.814 "read": true, 00:10:10.814 "write": true, 00:10:10.814 "unmap": true, 00:10:10.814 "flush": true, 00:10:10.814 "reset": true, 00:10:10.814 "nvme_admin": false, 00:10:10.814 "nvme_io": false, 00:10:10.814 "nvme_io_md": false, 00:10:10.814 "write_zeroes": true, 00:10:10.814 "zcopy": true, 00:10:10.814 "get_zone_info": false, 00:10:10.814 "zone_management": false, 00:10:10.814 "zone_append": false, 00:10:10.814 "compare": false, 00:10:10.814 "compare_and_write": false, 00:10:10.814 "abort": true, 00:10:10.814 "seek_hole": false, 00:10:10.814 "seek_data": false, 00:10:10.814 "copy": true, 00:10:10.814 "nvme_iov_md": false 00:10:10.814 }, 00:10:10.814 "memory_domains": [ 00:10:10.814 { 00:10:10.814 "dma_device_id": "system", 00:10:10.814 "dma_device_type": 1 00:10:10.814 }, 00:10:10.814 { 00:10:10.814 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:10.814 "dma_device_type": 2 00:10:10.814 } 00:10:10.814 ], 00:10:10.814 "driver_specific": {} 00:10:10.814 } 00:10:10.814 ] 00:10:10.814 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.814 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:10.814 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:10.814 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:10.815 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:10.815 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:10.815 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:10.815 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:10.815 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:10.815 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:10.815 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:10.815 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:10.815 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:10.815 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:10.815 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:10.815 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:10.815 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.815 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.073 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.073 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:11.074 "name": "Existed_Raid", 00:10:11.074 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:11.074 "strip_size_kb": 64, 00:10:11.074 "state": "configuring", 00:10:11.074 "raid_level": "concat", 00:10:11.074 "superblock": false, 00:10:11.074 "num_base_bdevs": 4, 00:10:11.074 "num_base_bdevs_discovered": 2, 00:10:11.074 "num_base_bdevs_operational": 4, 00:10:11.074 "base_bdevs_list": [ 00:10:11.074 { 00:10:11.074 "name": "BaseBdev1", 00:10:11.074 "uuid": "10d0ce0e-51ff-476f-b148-187393f14178", 00:10:11.074 "is_configured": true, 00:10:11.074 "data_offset": 0, 00:10:11.074 "data_size": 65536 00:10:11.074 }, 00:10:11.074 { 00:10:11.074 "name": "BaseBdev2", 00:10:11.074 "uuid": "357aadef-565c-49b4-9313-72cf3da4824b", 00:10:11.074 "is_configured": true, 00:10:11.074 "data_offset": 0, 00:10:11.074 "data_size": 65536 00:10:11.074 }, 00:10:11.074 { 00:10:11.074 "name": "BaseBdev3", 00:10:11.074 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:11.074 "is_configured": false, 00:10:11.074 "data_offset": 0, 00:10:11.074 "data_size": 0 00:10:11.074 }, 00:10:11.074 { 00:10:11.074 "name": "BaseBdev4", 00:10:11.074 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:11.074 "is_configured": false, 00:10:11.074 "data_offset": 0, 00:10:11.074 "data_size": 0 00:10:11.074 } 00:10:11.074 ] 00:10:11.074 }' 00:10:11.074 15:14:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:11.074 15:14:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.331 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:11.331 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.331 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.331 [2024-11-27 15:14:39.357390] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:11.331 BaseBdev3 00:10:11.331 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.331 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:10:11.331 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:11.331 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:11.331 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:11.331 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:11.331 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:11.331 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:11.331 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.331 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.331 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.331 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:11.331 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.331 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.331 [ 00:10:11.331 { 00:10:11.331 "name": "BaseBdev3", 00:10:11.331 "aliases": [ 00:10:11.331 "d6fcd938-d447-405a-a527-9a8ac232e153" 00:10:11.331 ], 00:10:11.331 "product_name": "Malloc disk", 00:10:11.331 "block_size": 512, 00:10:11.331 "num_blocks": 65536, 00:10:11.331 "uuid": "d6fcd938-d447-405a-a527-9a8ac232e153", 00:10:11.331 "assigned_rate_limits": { 00:10:11.331 "rw_ios_per_sec": 0, 00:10:11.331 "rw_mbytes_per_sec": 0, 00:10:11.331 "r_mbytes_per_sec": 0, 00:10:11.331 "w_mbytes_per_sec": 0 00:10:11.331 }, 00:10:11.331 "claimed": true, 00:10:11.331 "claim_type": "exclusive_write", 00:10:11.331 "zoned": false, 00:10:11.331 "supported_io_types": { 00:10:11.331 "read": true, 00:10:11.331 "write": true, 00:10:11.331 "unmap": true, 00:10:11.331 "flush": true, 00:10:11.331 "reset": true, 00:10:11.331 "nvme_admin": false, 00:10:11.331 "nvme_io": false, 00:10:11.331 "nvme_io_md": false, 00:10:11.331 "write_zeroes": true, 00:10:11.331 "zcopy": true, 00:10:11.331 "get_zone_info": false, 00:10:11.331 "zone_management": false, 00:10:11.331 "zone_append": false, 00:10:11.331 "compare": false, 00:10:11.331 "compare_and_write": false, 00:10:11.331 "abort": true, 00:10:11.331 "seek_hole": false, 00:10:11.331 "seek_data": false, 00:10:11.331 "copy": true, 00:10:11.331 "nvme_iov_md": false 00:10:11.331 }, 00:10:11.331 "memory_domains": [ 00:10:11.331 { 00:10:11.331 "dma_device_id": "system", 00:10:11.331 "dma_device_type": 1 00:10:11.331 }, 00:10:11.331 { 00:10:11.331 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:11.331 "dma_device_type": 2 00:10:11.332 } 00:10:11.332 ], 00:10:11.332 "driver_specific": {} 00:10:11.332 } 00:10:11.332 ] 00:10:11.332 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.332 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:11.332 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:11.332 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:11.332 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:11.332 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:11.332 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:11.332 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:11.332 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:11.332 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:11.332 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:11.332 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:11.332 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:11.332 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:11.332 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:11.332 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:11.332 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.332 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.332 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.590 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:11.590 "name": "Existed_Raid", 00:10:11.590 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:11.590 "strip_size_kb": 64, 00:10:11.590 "state": "configuring", 00:10:11.590 "raid_level": "concat", 00:10:11.590 "superblock": false, 00:10:11.590 "num_base_bdevs": 4, 00:10:11.590 "num_base_bdevs_discovered": 3, 00:10:11.590 "num_base_bdevs_operational": 4, 00:10:11.590 "base_bdevs_list": [ 00:10:11.590 { 00:10:11.590 "name": "BaseBdev1", 00:10:11.590 "uuid": "10d0ce0e-51ff-476f-b148-187393f14178", 00:10:11.590 "is_configured": true, 00:10:11.590 "data_offset": 0, 00:10:11.590 "data_size": 65536 00:10:11.590 }, 00:10:11.590 { 00:10:11.590 "name": "BaseBdev2", 00:10:11.590 "uuid": "357aadef-565c-49b4-9313-72cf3da4824b", 00:10:11.590 "is_configured": true, 00:10:11.590 "data_offset": 0, 00:10:11.590 "data_size": 65536 00:10:11.590 }, 00:10:11.590 { 00:10:11.590 "name": "BaseBdev3", 00:10:11.590 "uuid": "d6fcd938-d447-405a-a527-9a8ac232e153", 00:10:11.590 "is_configured": true, 00:10:11.590 "data_offset": 0, 00:10:11.590 "data_size": 65536 00:10:11.590 }, 00:10:11.590 { 00:10:11.590 "name": "BaseBdev4", 00:10:11.590 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:11.590 "is_configured": false, 00:10:11.590 "data_offset": 0, 00:10:11.590 "data_size": 0 00:10:11.590 } 00:10:11.590 ] 00:10:11.590 }' 00:10:11.590 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:11.590 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.848 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:11.848 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.848 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.848 [2024-11-27 15:14:39.755738] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:11.848 [2024-11-27 15:14:39.755784] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:10:11.848 [2024-11-27 15:14:39.755792] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:10:11.848 [2024-11-27 15:14:39.756106] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:10:11.848 [2024-11-27 15:14:39.756257] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:10:11.848 [2024-11-27 15:14:39.756270] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:10:11.848 [2024-11-27 15:14:39.756477] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:11.848 BaseBdev4 00:10:11.848 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.848 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:10:11.848 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:10:11.848 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:11.848 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:11.848 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:11.848 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:11.848 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:11.848 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.848 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.848 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.848 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:11.848 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.848 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.848 [ 00:10:11.848 { 00:10:11.848 "name": "BaseBdev4", 00:10:11.848 "aliases": [ 00:10:11.848 "d7b2c820-1d1f-473e-8fa3-e4c228022a06" 00:10:11.848 ], 00:10:11.848 "product_name": "Malloc disk", 00:10:11.848 "block_size": 512, 00:10:11.848 "num_blocks": 65536, 00:10:11.848 "uuid": "d7b2c820-1d1f-473e-8fa3-e4c228022a06", 00:10:11.848 "assigned_rate_limits": { 00:10:11.848 "rw_ios_per_sec": 0, 00:10:11.848 "rw_mbytes_per_sec": 0, 00:10:11.848 "r_mbytes_per_sec": 0, 00:10:11.848 "w_mbytes_per_sec": 0 00:10:11.848 }, 00:10:11.848 "claimed": true, 00:10:11.848 "claim_type": "exclusive_write", 00:10:11.848 "zoned": false, 00:10:11.848 "supported_io_types": { 00:10:11.848 "read": true, 00:10:11.848 "write": true, 00:10:11.848 "unmap": true, 00:10:11.848 "flush": true, 00:10:11.848 "reset": true, 00:10:11.848 "nvme_admin": false, 00:10:11.848 "nvme_io": false, 00:10:11.848 "nvme_io_md": false, 00:10:11.848 "write_zeroes": true, 00:10:11.848 "zcopy": true, 00:10:11.848 "get_zone_info": false, 00:10:11.848 "zone_management": false, 00:10:11.848 "zone_append": false, 00:10:11.848 "compare": false, 00:10:11.848 "compare_and_write": false, 00:10:11.848 "abort": true, 00:10:11.848 "seek_hole": false, 00:10:11.848 "seek_data": false, 00:10:11.848 "copy": true, 00:10:11.848 "nvme_iov_md": false 00:10:11.848 }, 00:10:11.848 "memory_domains": [ 00:10:11.848 { 00:10:11.848 "dma_device_id": "system", 00:10:11.848 "dma_device_type": 1 00:10:11.848 }, 00:10:11.848 { 00:10:11.848 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:11.848 "dma_device_type": 2 00:10:11.848 } 00:10:11.848 ], 00:10:11.848 "driver_specific": {} 00:10:11.848 } 00:10:11.848 ] 00:10:11.848 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.848 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:11.848 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:11.848 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:11.848 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:10:11.848 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:11.848 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:11.849 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:11.849 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:11.849 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:11.849 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:11.849 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:11.849 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:11.849 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:11.849 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:11.849 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.849 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:11.849 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.849 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.849 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:11.849 "name": "Existed_Raid", 00:10:11.849 "uuid": "7a3e8b0c-d13f-4a18-a08e-676e14860f95", 00:10:11.849 "strip_size_kb": 64, 00:10:11.849 "state": "online", 00:10:11.849 "raid_level": "concat", 00:10:11.849 "superblock": false, 00:10:11.849 "num_base_bdevs": 4, 00:10:11.849 "num_base_bdevs_discovered": 4, 00:10:11.849 "num_base_bdevs_operational": 4, 00:10:11.849 "base_bdevs_list": [ 00:10:11.849 { 00:10:11.849 "name": "BaseBdev1", 00:10:11.849 "uuid": "10d0ce0e-51ff-476f-b148-187393f14178", 00:10:11.849 "is_configured": true, 00:10:11.849 "data_offset": 0, 00:10:11.849 "data_size": 65536 00:10:11.849 }, 00:10:11.849 { 00:10:11.849 "name": "BaseBdev2", 00:10:11.849 "uuid": "357aadef-565c-49b4-9313-72cf3da4824b", 00:10:11.849 "is_configured": true, 00:10:11.849 "data_offset": 0, 00:10:11.849 "data_size": 65536 00:10:11.849 }, 00:10:11.849 { 00:10:11.849 "name": "BaseBdev3", 00:10:11.849 "uuid": "d6fcd938-d447-405a-a527-9a8ac232e153", 00:10:11.849 "is_configured": true, 00:10:11.849 "data_offset": 0, 00:10:11.849 "data_size": 65536 00:10:11.849 }, 00:10:11.849 { 00:10:11.849 "name": "BaseBdev4", 00:10:11.849 "uuid": "d7b2c820-1d1f-473e-8fa3-e4c228022a06", 00:10:11.849 "is_configured": true, 00:10:11.849 "data_offset": 0, 00:10:11.849 "data_size": 65536 00:10:11.849 } 00:10:11.849 ] 00:10:11.849 }' 00:10:11.849 15:14:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:11.849 15:14:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.108 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:12.108 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:12.108 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:12.108 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:12.108 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:12.108 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:12.108 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:12.108 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:12.108 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.108 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:12.108 [2024-11-27 15:14:40.171441] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:12.108 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:12.108 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:12.108 "name": "Existed_Raid", 00:10:12.108 "aliases": [ 00:10:12.108 "7a3e8b0c-d13f-4a18-a08e-676e14860f95" 00:10:12.108 ], 00:10:12.108 "product_name": "Raid Volume", 00:10:12.108 "block_size": 512, 00:10:12.108 "num_blocks": 262144, 00:10:12.108 "uuid": "7a3e8b0c-d13f-4a18-a08e-676e14860f95", 00:10:12.108 "assigned_rate_limits": { 00:10:12.108 "rw_ios_per_sec": 0, 00:10:12.108 "rw_mbytes_per_sec": 0, 00:10:12.108 "r_mbytes_per_sec": 0, 00:10:12.108 "w_mbytes_per_sec": 0 00:10:12.108 }, 00:10:12.108 "claimed": false, 00:10:12.108 "zoned": false, 00:10:12.108 "supported_io_types": { 00:10:12.108 "read": true, 00:10:12.108 "write": true, 00:10:12.108 "unmap": true, 00:10:12.108 "flush": true, 00:10:12.108 "reset": true, 00:10:12.108 "nvme_admin": false, 00:10:12.108 "nvme_io": false, 00:10:12.108 "nvme_io_md": false, 00:10:12.108 "write_zeroes": true, 00:10:12.108 "zcopy": false, 00:10:12.108 "get_zone_info": false, 00:10:12.108 "zone_management": false, 00:10:12.108 "zone_append": false, 00:10:12.108 "compare": false, 00:10:12.108 "compare_and_write": false, 00:10:12.108 "abort": false, 00:10:12.108 "seek_hole": false, 00:10:12.108 "seek_data": false, 00:10:12.108 "copy": false, 00:10:12.108 "nvme_iov_md": false 00:10:12.108 }, 00:10:12.108 "memory_domains": [ 00:10:12.108 { 00:10:12.108 "dma_device_id": "system", 00:10:12.108 "dma_device_type": 1 00:10:12.108 }, 00:10:12.108 { 00:10:12.108 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:12.108 "dma_device_type": 2 00:10:12.108 }, 00:10:12.108 { 00:10:12.108 "dma_device_id": "system", 00:10:12.108 "dma_device_type": 1 00:10:12.108 }, 00:10:12.108 { 00:10:12.108 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:12.108 "dma_device_type": 2 00:10:12.108 }, 00:10:12.108 { 00:10:12.108 "dma_device_id": "system", 00:10:12.108 "dma_device_type": 1 00:10:12.108 }, 00:10:12.108 { 00:10:12.108 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:12.108 "dma_device_type": 2 00:10:12.108 }, 00:10:12.108 { 00:10:12.108 "dma_device_id": "system", 00:10:12.108 "dma_device_type": 1 00:10:12.108 }, 00:10:12.108 { 00:10:12.108 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:12.108 "dma_device_type": 2 00:10:12.108 } 00:10:12.108 ], 00:10:12.108 "driver_specific": { 00:10:12.108 "raid": { 00:10:12.108 "uuid": "7a3e8b0c-d13f-4a18-a08e-676e14860f95", 00:10:12.108 "strip_size_kb": 64, 00:10:12.108 "state": "online", 00:10:12.108 "raid_level": "concat", 00:10:12.108 "superblock": false, 00:10:12.108 "num_base_bdevs": 4, 00:10:12.108 "num_base_bdevs_discovered": 4, 00:10:12.108 "num_base_bdevs_operational": 4, 00:10:12.108 "base_bdevs_list": [ 00:10:12.108 { 00:10:12.108 "name": "BaseBdev1", 00:10:12.108 "uuid": "10d0ce0e-51ff-476f-b148-187393f14178", 00:10:12.108 "is_configured": true, 00:10:12.108 "data_offset": 0, 00:10:12.108 "data_size": 65536 00:10:12.108 }, 00:10:12.108 { 00:10:12.108 "name": "BaseBdev2", 00:10:12.108 "uuid": "357aadef-565c-49b4-9313-72cf3da4824b", 00:10:12.108 "is_configured": true, 00:10:12.108 "data_offset": 0, 00:10:12.108 "data_size": 65536 00:10:12.108 }, 00:10:12.108 { 00:10:12.108 "name": "BaseBdev3", 00:10:12.108 "uuid": "d6fcd938-d447-405a-a527-9a8ac232e153", 00:10:12.109 "is_configured": true, 00:10:12.109 "data_offset": 0, 00:10:12.109 "data_size": 65536 00:10:12.109 }, 00:10:12.109 { 00:10:12.109 "name": "BaseBdev4", 00:10:12.109 "uuid": "d7b2c820-1d1f-473e-8fa3-e4c228022a06", 00:10:12.109 "is_configured": true, 00:10:12.109 "data_offset": 0, 00:10:12.109 "data_size": 65536 00:10:12.109 } 00:10:12.109 ] 00:10:12.109 } 00:10:12.109 } 00:10:12.109 }' 00:10:12.109 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:12.368 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:12.368 BaseBdev2 00:10:12.368 BaseBdev3 00:10:12.368 BaseBdev4' 00:10:12.368 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:12.368 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:12.368 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:12.368 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:12.368 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:12.368 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:12.368 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.368 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:12.368 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:12.368 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:12.368 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:12.368 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:12.368 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:12.368 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.368 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:12.368 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:12.368 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:12.368 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:12.368 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:12.369 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:12.369 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:12.369 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:12.369 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.369 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:12.369 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:12.369 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:12.369 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:12.369 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:12.369 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:12.369 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:12.369 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.369 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:12.369 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:12.369 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:12.369 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:12.369 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:12.369 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.369 [2024-11-27 15:14:40.458707] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:12.369 [2024-11-27 15:14:40.458744] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:12.369 [2024-11-27 15:14:40.458804] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:12.628 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:12.628 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:12.628 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:10:12.628 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:12.628 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:12.628 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:10:12.628 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 3 00:10:12.628 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:12.628 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:10:12.628 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:12.628 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:12.628 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:12.628 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:12.628 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:12.628 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:12.628 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:12.628 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:12.628 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:12.628 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:12.628 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.628 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:12.628 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:12.628 "name": "Existed_Raid", 00:10:12.628 "uuid": "7a3e8b0c-d13f-4a18-a08e-676e14860f95", 00:10:12.628 "strip_size_kb": 64, 00:10:12.628 "state": "offline", 00:10:12.628 "raid_level": "concat", 00:10:12.628 "superblock": false, 00:10:12.628 "num_base_bdevs": 4, 00:10:12.628 "num_base_bdevs_discovered": 3, 00:10:12.628 "num_base_bdevs_operational": 3, 00:10:12.628 "base_bdevs_list": [ 00:10:12.628 { 00:10:12.628 "name": null, 00:10:12.628 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:12.628 "is_configured": false, 00:10:12.628 "data_offset": 0, 00:10:12.628 "data_size": 65536 00:10:12.628 }, 00:10:12.628 { 00:10:12.628 "name": "BaseBdev2", 00:10:12.628 "uuid": "357aadef-565c-49b4-9313-72cf3da4824b", 00:10:12.628 "is_configured": true, 00:10:12.628 "data_offset": 0, 00:10:12.628 "data_size": 65536 00:10:12.628 }, 00:10:12.629 { 00:10:12.629 "name": "BaseBdev3", 00:10:12.629 "uuid": "d6fcd938-d447-405a-a527-9a8ac232e153", 00:10:12.629 "is_configured": true, 00:10:12.629 "data_offset": 0, 00:10:12.629 "data_size": 65536 00:10:12.629 }, 00:10:12.629 { 00:10:12.629 "name": "BaseBdev4", 00:10:12.629 "uuid": "d7b2c820-1d1f-473e-8fa3-e4c228022a06", 00:10:12.629 "is_configured": true, 00:10:12.629 "data_offset": 0, 00:10:12.629 "data_size": 65536 00:10:12.629 } 00:10:12.629 ] 00:10:12.629 }' 00:10:12.629 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:12.629 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.888 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:12.888 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:12.888 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:12.888 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:12.888 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:12.888 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.888 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:12.888 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:12.888 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:12.888 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:12.888 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:12.888 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.888 [2024-11-27 15:14:40.925376] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:12.888 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:12.888 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:12.888 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:12.888 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:12.888 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:12.888 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.888 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:12.888 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:12.888 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:12.888 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:12.888 15:14:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:12.888 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:12.888 15:14:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.148 [2024-11-27 15:14:40.992647] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:13.148 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.148 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:13.148 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:13.148 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:13.148 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:13.148 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.148 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.148 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.148 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:13.148 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:13.148 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:10:13.148 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.148 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.148 [2024-11-27 15:14:41.055850] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:10:13.148 [2024-11-27 15:14:41.055955] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:10:13.148 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.148 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.149 BaseBdev2 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.149 [ 00:10:13.149 { 00:10:13.149 "name": "BaseBdev2", 00:10:13.149 "aliases": [ 00:10:13.149 "2d09c754-4550-49ac-ab3d-4738e4e9a3cc" 00:10:13.149 ], 00:10:13.149 "product_name": "Malloc disk", 00:10:13.149 "block_size": 512, 00:10:13.149 "num_blocks": 65536, 00:10:13.149 "uuid": "2d09c754-4550-49ac-ab3d-4738e4e9a3cc", 00:10:13.149 "assigned_rate_limits": { 00:10:13.149 "rw_ios_per_sec": 0, 00:10:13.149 "rw_mbytes_per_sec": 0, 00:10:13.149 "r_mbytes_per_sec": 0, 00:10:13.149 "w_mbytes_per_sec": 0 00:10:13.149 }, 00:10:13.149 "claimed": false, 00:10:13.149 "zoned": false, 00:10:13.149 "supported_io_types": { 00:10:13.149 "read": true, 00:10:13.149 "write": true, 00:10:13.149 "unmap": true, 00:10:13.149 "flush": true, 00:10:13.149 "reset": true, 00:10:13.149 "nvme_admin": false, 00:10:13.149 "nvme_io": false, 00:10:13.149 "nvme_io_md": false, 00:10:13.149 "write_zeroes": true, 00:10:13.149 "zcopy": true, 00:10:13.149 "get_zone_info": false, 00:10:13.149 "zone_management": false, 00:10:13.149 "zone_append": false, 00:10:13.149 "compare": false, 00:10:13.149 "compare_and_write": false, 00:10:13.149 "abort": true, 00:10:13.149 "seek_hole": false, 00:10:13.149 "seek_data": false, 00:10:13.149 "copy": true, 00:10:13.149 "nvme_iov_md": false 00:10:13.149 }, 00:10:13.149 "memory_domains": [ 00:10:13.149 { 00:10:13.149 "dma_device_id": "system", 00:10:13.149 "dma_device_type": 1 00:10:13.149 }, 00:10:13.149 { 00:10:13.149 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:13.149 "dma_device_type": 2 00:10:13.149 } 00:10:13.149 ], 00:10:13.149 "driver_specific": {} 00:10:13.149 } 00:10:13.149 ] 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.149 BaseBdev3 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.149 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.149 [ 00:10:13.149 { 00:10:13.149 "name": "BaseBdev3", 00:10:13.149 "aliases": [ 00:10:13.149 "dff4c5e3-ab11-48ca-974d-fa6ed16fedb6" 00:10:13.149 ], 00:10:13.149 "product_name": "Malloc disk", 00:10:13.149 "block_size": 512, 00:10:13.149 "num_blocks": 65536, 00:10:13.149 "uuid": "dff4c5e3-ab11-48ca-974d-fa6ed16fedb6", 00:10:13.149 "assigned_rate_limits": { 00:10:13.149 "rw_ios_per_sec": 0, 00:10:13.149 "rw_mbytes_per_sec": 0, 00:10:13.149 "r_mbytes_per_sec": 0, 00:10:13.149 "w_mbytes_per_sec": 0 00:10:13.149 }, 00:10:13.149 "claimed": false, 00:10:13.149 "zoned": false, 00:10:13.149 "supported_io_types": { 00:10:13.149 "read": true, 00:10:13.149 "write": true, 00:10:13.149 "unmap": true, 00:10:13.149 "flush": true, 00:10:13.149 "reset": true, 00:10:13.149 "nvme_admin": false, 00:10:13.149 "nvme_io": false, 00:10:13.149 "nvme_io_md": false, 00:10:13.149 "write_zeroes": true, 00:10:13.149 "zcopy": true, 00:10:13.149 "get_zone_info": false, 00:10:13.149 "zone_management": false, 00:10:13.149 "zone_append": false, 00:10:13.149 "compare": false, 00:10:13.150 "compare_and_write": false, 00:10:13.150 "abort": true, 00:10:13.150 "seek_hole": false, 00:10:13.150 "seek_data": false, 00:10:13.150 "copy": true, 00:10:13.150 "nvme_iov_md": false 00:10:13.150 }, 00:10:13.150 "memory_domains": [ 00:10:13.150 { 00:10:13.150 "dma_device_id": "system", 00:10:13.150 "dma_device_type": 1 00:10:13.150 }, 00:10:13.150 { 00:10:13.150 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:13.150 "dma_device_type": 2 00:10:13.150 } 00:10:13.150 ], 00:10:13.150 "driver_specific": {} 00:10:13.150 } 00:10:13.150 ] 00:10:13.150 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.150 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:13.150 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:13.150 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:13.150 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:13.150 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.150 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.150 BaseBdev4 00:10:13.150 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.150 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:10:13.150 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:10:13.150 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:13.150 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:13.150 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:13.150 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:13.150 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:13.150 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.150 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.409 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.409 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:13.409 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.409 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.409 [ 00:10:13.409 { 00:10:13.409 "name": "BaseBdev4", 00:10:13.409 "aliases": [ 00:10:13.409 "8f8d6298-3fc1-4c2f-929f-c741134176c9" 00:10:13.409 ], 00:10:13.409 "product_name": "Malloc disk", 00:10:13.409 "block_size": 512, 00:10:13.409 "num_blocks": 65536, 00:10:13.409 "uuid": "8f8d6298-3fc1-4c2f-929f-c741134176c9", 00:10:13.409 "assigned_rate_limits": { 00:10:13.409 "rw_ios_per_sec": 0, 00:10:13.409 "rw_mbytes_per_sec": 0, 00:10:13.409 "r_mbytes_per_sec": 0, 00:10:13.409 "w_mbytes_per_sec": 0 00:10:13.409 }, 00:10:13.409 "claimed": false, 00:10:13.409 "zoned": false, 00:10:13.409 "supported_io_types": { 00:10:13.409 "read": true, 00:10:13.409 "write": true, 00:10:13.409 "unmap": true, 00:10:13.409 "flush": true, 00:10:13.409 "reset": true, 00:10:13.409 "nvme_admin": false, 00:10:13.409 "nvme_io": false, 00:10:13.409 "nvme_io_md": false, 00:10:13.409 "write_zeroes": true, 00:10:13.409 "zcopy": true, 00:10:13.409 "get_zone_info": false, 00:10:13.409 "zone_management": false, 00:10:13.409 "zone_append": false, 00:10:13.409 "compare": false, 00:10:13.409 "compare_and_write": false, 00:10:13.409 "abort": true, 00:10:13.409 "seek_hole": false, 00:10:13.409 "seek_data": false, 00:10:13.409 "copy": true, 00:10:13.409 "nvme_iov_md": false 00:10:13.409 }, 00:10:13.409 "memory_domains": [ 00:10:13.409 { 00:10:13.409 "dma_device_id": "system", 00:10:13.409 "dma_device_type": 1 00:10:13.409 }, 00:10:13.409 { 00:10:13.409 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:13.409 "dma_device_type": 2 00:10:13.409 } 00:10:13.409 ], 00:10:13.409 "driver_specific": {} 00:10:13.409 } 00:10:13.409 ] 00:10:13.409 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.409 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:13.409 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:13.409 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:13.409 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:13.409 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.409 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.409 [2024-11-27 15:14:41.289010] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:13.409 [2024-11-27 15:14:41.289056] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:13.409 [2024-11-27 15:14:41.289078] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:13.409 [2024-11-27 15:14:41.290862] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:13.409 [2024-11-27 15:14:41.290926] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:13.409 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.409 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:13.409 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:13.409 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:13.409 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:13.409 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:13.409 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:13.409 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:13.409 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:13.409 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:13.409 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:13.409 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:13.409 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.409 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:13.409 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.409 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.409 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:13.409 "name": "Existed_Raid", 00:10:13.409 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:13.409 "strip_size_kb": 64, 00:10:13.409 "state": "configuring", 00:10:13.409 "raid_level": "concat", 00:10:13.409 "superblock": false, 00:10:13.409 "num_base_bdevs": 4, 00:10:13.409 "num_base_bdevs_discovered": 3, 00:10:13.409 "num_base_bdevs_operational": 4, 00:10:13.409 "base_bdevs_list": [ 00:10:13.409 { 00:10:13.410 "name": "BaseBdev1", 00:10:13.410 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:13.410 "is_configured": false, 00:10:13.410 "data_offset": 0, 00:10:13.410 "data_size": 0 00:10:13.410 }, 00:10:13.410 { 00:10:13.410 "name": "BaseBdev2", 00:10:13.410 "uuid": "2d09c754-4550-49ac-ab3d-4738e4e9a3cc", 00:10:13.410 "is_configured": true, 00:10:13.410 "data_offset": 0, 00:10:13.410 "data_size": 65536 00:10:13.410 }, 00:10:13.410 { 00:10:13.410 "name": "BaseBdev3", 00:10:13.410 "uuid": "dff4c5e3-ab11-48ca-974d-fa6ed16fedb6", 00:10:13.410 "is_configured": true, 00:10:13.410 "data_offset": 0, 00:10:13.410 "data_size": 65536 00:10:13.410 }, 00:10:13.410 { 00:10:13.410 "name": "BaseBdev4", 00:10:13.410 "uuid": "8f8d6298-3fc1-4c2f-929f-c741134176c9", 00:10:13.410 "is_configured": true, 00:10:13.410 "data_offset": 0, 00:10:13.410 "data_size": 65536 00:10:13.410 } 00:10:13.410 ] 00:10:13.410 }' 00:10:13.410 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:13.410 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.669 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:13.669 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.669 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.669 [2024-11-27 15:14:41.748224] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:13.669 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.669 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:13.669 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:13.669 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:13.669 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:13.669 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:13.669 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:13.669 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:13.669 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:13.669 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:13.669 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:13.669 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:13.669 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.669 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.669 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:13.928 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.928 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:13.928 "name": "Existed_Raid", 00:10:13.928 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:13.928 "strip_size_kb": 64, 00:10:13.928 "state": "configuring", 00:10:13.928 "raid_level": "concat", 00:10:13.928 "superblock": false, 00:10:13.928 "num_base_bdevs": 4, 00:10:13.928 "num_base_bdevs_discovered": 2, 00:10:13.928 "num_base_bdevs_operational": 4, 00:10:13.928 "base_bdevs_list": [ 00:10:13.928 { 00:10:13.928 "name": "BaseBdev1", 00:10:13.928 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:13.928 "is_configured": false, 00:10:13.928 "data_offset": 0, 00:10:13.928 "data_size": 0 00:10:13.928 }, 00:10:13.928 { 00:10:13.928 "name": null, 00:10:13.928 "uuid": "2d09c754-4550-49ac-ab3d-4738e4e9a3cc", 00:10:13.928 "is_configured": false, 00:10:13.928 "data_offset": 0, 00:10:13.928 "data_size": 65536 00:10:13.928 }, 00:10:13.928 { 00:10:13.928 "name": "BaseBdev3", 00:10:13.928 "uuid": "dff4c5e3-ab11-48ca-974d-fa6ed16fedb6", 00:10:13.928 "is_configured": true, 00:10:13.928 "data_offset": 0, 00:10:13.928 "data_size": 65536 00:10:13.928 }, 00:10:13.928 { 00:10:13.928 "name": "BaseBdev4", 00:10:13.928 "uuid": "8f8d6298-3fc1-4c2f-929f-c741134176c9", 00:10:13.929 "is_configured": true, 00:10:13.929 "data_offset": 0, 00:10:13.929 "data_size": 65536 00:10:13.929 } 00:10:13.929 ] 00:10:13.929 }' 00:10:13.929 15:14:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:13.929 15:14:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.188 [2024-11-27 15:14:42.206498] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:14.188 BaseBdev1 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.188 [ 00:10:14.188 { 00:10:14.188 "name": "BaseBdev1", 00:10:14.188 "aliases": [ 00:10:14.188 "3172f577-9eb3-4141-9cee-19ef20e5e166" 00:10:14.188 ], 00:10:14.188 "product_name": "Malloc disk", 00:10:14.188 "block_size": 512, 00:10:14.188 "num_blocks": 65536, 00:10:14.188 "uuid": "3172f577-9eb3-4141-9cee-19ef20e5e166", 00:10:14.188 "assigned_rate_limits": { 00:10:14.188 "rw_ios_per_sec": 0, 00:10:14.188 "rw_mbytes_per_sec": 0, 00:10:14.188 "r_mbytes_per_sec": 0, 00:10:14.188 "w_mbytes_per_sec": 0 00:10:14.188 }, 00:10:14.188 "claimed": true, 00:10:14.188 "claim_type": "exclusive_write", 00:10:14.188 "zoned": false, 00:10:14.188 "supported_io_types": { 00:10:14.188 "read": true, 00:10:14.188 "write": true, 00:10:14.188 "unmap": true, 00:10:14.188 "flush": true, 00:10:14.188 "reset": true, 00:10:14.188 "nvme_admin": false, 00:10:14.188 "nvme_io": false, 00:10:14.188 "nvme_io_md": false, 00:10:14.188 "write_zeroes": true, 00:10:14.188 "zcopy": true, 00:10:14.188 "get_zone_info": false, 00:10:14.188 "zone_management": false, 00:10:14.188 "zone_append": false, 00:10:14.188 "compare": false, 00:10:14.188 "compare_and_write": false, 00:10:14.188 "abort": true, 00:10:14.188 "seek_hole": false, 00:10:14.188 "seek_data": false, 00:10:14.188 "copy": true, 00:10:14.188 "nvme_iov_md": false 00:10:14.188 }, 00:10:14.188 "memory_domains": [ 00:10:14.188 { 00:10:14.188 "dma_device_id": "system", 00:10:14.188 "dma_device_type": 1 00:10:14.188 }, 00:10:14.188 { 00:10:14.188 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:14.188 "dma_device_type": 2 00:10:14.188 } 00:10:14.188 ], 00:10:14.188 "driver_specific": {} 00:10:14.188 } 00:10:14.188 ] 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.188 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:14.189 "name": "Existed_Raid", 00:10:14.189 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:14.189 "strip_size_kb": 64, 00:10:14.189 "state": "configuring", 00:10:14.189 "raid_level": "concat", 00:10:14.189 "superblock": false, 00:10:14.189 "num_base_bdevs": 4, 00:10:14.189 "num_base_bdevs_discovered": 3, 00:10:14.189 "num_base_bdevs_operational": 4, 00:10:14.189 "base_bdevs_list": [ 00:10:14.189 { 00:10:14.189 "name": "BaseBdev1", 00:10:14.189 "uuid": "3172f577-9eb3-4141-9cee-19ef20e5e166", 00:10:14.189 "is_configured": true, 00:10:14.189 "data_offset": 0, 00:10:14.189 "data_size": 65536 00:10:14.189 }, 00:10:14.189 { 00:10:14.189 "name": null, 00:10:14.189 "uuid": "2d09c754-4550-49ac-ab3d-4738e4e9a3cc", 00:10:14.189 "is_configured": false, 00:10:14.189 "data_offset": 0, 00:10:14.189 "data_size": 65536 00:10:14.189 }, 00:10:14.189 { 00:10:14.189 "name": "BaseBdev3", 00:10:14.189 "uuid": "dff4c5e3-ab11-48ca-974d-fa6ed16fedb6", 00:10:14.189 "is_configured": true, 00:10:14.189 "data_offset": 0, 00:10:14.189 "data_size": 65536 00:10:14.189 }, 00:10:14.189 { 00:10:14.189 "name": "BaseBdev4", 00:10:14.189 "uuid": "8f8d6298-3fc1-4c2f-929f-c741134176c9", 00:10:14.189 "is_configured": true, 00:10:14.189 "data_offset": 0, 00:10:14.189 "data_size": 65536 00:10:14.189 } 00:10:14.189 ] 00:10:14.189 }' 00:10:14.189 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:14.189 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.756 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:14.756 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.756 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.756 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:14.756 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.756 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:14.756 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:14.756 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.756 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.756 [2024-11-27 15:14:42.685778] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:14.756 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.756 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:14.756 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:14.756 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:14.756 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:14.756 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:14.756 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:14.756 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:14.756 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:14.756 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:14.756 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:14.756 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:14.756 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:14.756 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.756 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.756 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.756 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:14.756 "name": "Existed_Raid", 00:10:14.756 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:14.756 "strip_size_kb": 64, 00:10:14.756 "state": "configuring", 00:10:14.756 "raid_level": "concat", 00:10:14.756 "superblock": false, 00:10:14.756 "num_base_bdevs": 4, 00:10:14.756 "num_base_bdevs_discovered": 2, 00:10:14.756 "num_base_bdevs_operational": 4, 00:10:14.756 "base_bdevs_list": [ 00:10:14.756 { 00:10:14.756 "name": "BaseBdev1", 00:10:14.756 "uuid": "3172f577-9eb3-4141-9cee-19ef20e5e166", 00:10:14.756 "is_configured": true, 00:10:14.756 "data_offset": 0, 00:10:14.756 "data_size": 65536 00:10:14.756 }, 00:10:14.756 { 00:10:14.756 "name": null, 00:10:14.756 "uuid": "2d09c754-4550-49ac-ab3d-4738e4e9a3cc", 00:10:14.756 "is_configured": false, 00:10:14.756 "data_offset": 0, 00:10:14.756 "data_size": 65536 00:10:14.756 }, 00:10:14.756 { 00:10:14.756 "name": null, 00:10:14.756 "uuid": "dff4c5e3-ab11-48ca-974d-fa6ed16fedb6", 00:10:14.756 "is_configured": false, 00:10:14.756 "data_offset": 0, 00:10:14.756 "data_size": 65536 00:10:14.756 }, 00:10:14.756 { 00:10:14.756 "name": "BaseBdev4", 00:10:14.756 "uuid": "8f8d6298-3fc1-4c2f-929f-c741134176c9", 00:10:14.756 "is_configured": true, 00:10:14.756 "data_offset": 0, 00:10:14.756 "data_size": 65536 00:10:14.756 } 00:10:14.756 ] 00:10:14.756 }' 00:10:14.756 15:14:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:14.756 15:14:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.377 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:15.377 15:14:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.377 15:14:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.377 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:15.377 15:14:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.377 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:15.377 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:15.377 15:14:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.377 15:14:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.377 [2024-11-27 15:14:43.192916] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:15.377 15:14:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.377 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:15.377 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:15.377 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:15.377 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:15.377 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:15.377 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:15.377 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:15.377 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:15.377 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:15.377 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:15.377 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:15.377 15:14:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.377 15:14:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.377 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:15.377 15:14:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.377 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:15.377 "name": "Existed_Raid", 00:10:15.377 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:15.377 "strip_size_kb": 64, 00:10:15.377 "state": "configuring", 00:10:15.377 "raid_level": "concat", 00:10:15.377 "superblock": false, 00:10:15.377 "num_base_bdevs": 4, 00:10:15.377 "num_base_bdevs_discovered": 3, 00:10:15.377 "num_base_bdevs_operational": 4, 00:10:15.377 "base_bdevs_list": [ 00:10:15.377 { 00:10:15.377 "name": "BaseBdev1", 00:10:15.377 "uuid": "3172f577-9eb3-4141-9cee-19ef20e5e166", 00:10:15.377 "is_configured": true, 00:10:15.377 "data_offset": 0, 00:10:15.377 "data_size": 65536 00:10:15.377 }, 00:10:15.377 { 00:10:15.377 "name": null, 00:10:15.377 "uuid": "2d09c754-4550-49ac-ab3d-4738e4e9a3cc", 00:10:15.377 "is_configured": false, 00:10:15.377 "data_offset": 0, 00:10:15.377 "data_size": 65536 00:10:15.377 }, 00:10:15.377 { 00:10:15.377 "name": "BaseBdev3", 00:10:15.377 "uuid": "dff4c5e3-ab11-48ca-974d-fa6ed16fedb6", 00:10:15.377 "is_configured": true, 00:10:15.377 "data_offset": 0, 00:10:15.377 "data_size": 65536 00:10:15.377 }, 00:10:15.377 { 00:10:15.377 "name": "BaseBdev4", 00:10:15.377 "uuid": "8f8d6298-3fc1-4c2f-929f-c741134176c9", 00:10:15.377 "is_configured": true, 00:10:15.377 "data_offset": 0, 00:10:15.377 "data_size": 65536 00:10:15.377 } 00:10:15.377 ] 00:10:15.377 }' 00:10:15.377 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:15.377 15:14:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.636 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:15.636 15:14:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.636 15:14:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.636 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:15.636 15:14:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.636 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:15.636 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:15.636 15:14:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.636 15:14:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.636 [2024-11-27 15:14:43.656107] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:15.636 15:14:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.636 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:15.637 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:15.637 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:15.637 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:15.637 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:15.637 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:15.637 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:15.637 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:15.637 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:15.637 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:15.637 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:15.637 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:15.637 15:14:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.637 15:14:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.637 15:14:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.637 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:15.637 "name": "Existed_Raid", 00:10:15.637 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:15.637 "strip_size_kb": 64, 00:10:15.637 "state": "configuring", 00:10:15.637 "raid_level": "concat", 00:10:15.637 "superblock": false, 00:10:15.637 "num_base_bdevs": 4, 00:10:15.637 "num_base_bdevs_discovered": 2, 00:10:15.637 "num_base_bdevs_operational": 4, 00:10:15.637 "base_bdevs_list": [ 00:10:15.637 { 00:10:15.637 "name": null, 00:10:15.637 "uuid": "3172f577-9eb3-4141-9cee-19ef20e5e166", 00:10:15.637 "is_configured": false, 00:10:15.637 "data_offset": 0, 00:10:15.637 "data_size": 65536 00:10:15.637 }, 00:10:15.637 { 00:10:15.637 "name": null, 00:10:15.637 "uuid": "2d09c754-4550-49ac-ab3d-4738e4e9a3cc", 00:10:15.637 "is_configured": false, 00:10:15.637 "data_offset": 0, 00:10:15.637 "data_size": 65536 00:10:15.637 }, 00:10:15.637 { 00:10:15.637 "name": "BaseBdev3", 00:10:15.637 "uuid": "dff4c5e3-ab11-48ca-974d-fa6ed16fedb6", 00:10:15.637 "is_configured": true, 00:10:15.637 "data_offset": 0, 00:10:15.637 "data_size": 65536 00:10:15.637 }, 00:10:15.637 { 00:10:15.637 "name": "BaseBdev4", 00:10:15.637 "uuid": "8f8d6298-3fc1-4c2f-929f-c741134176c9", 00:10:15.637 "is_configured": true, 00:10:15.637 "data_offset": 0, 00:10:15.637 "data_size": 65536 00:10:15.637 } 00:10:15.637 ] 00:10:15.637 }' 00:10:15.637 15:14:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:15.637 15:14:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.205 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:16.205 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.205 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.205 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:16.205 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.205 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:16.205 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:16.205 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.205 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.205 [2024-11-27 15:14:44.109793] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:16.205 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.205 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:16.205 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:16.205 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:16.205 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:16.205 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:16.205 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:16.205 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:16.205 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:16.205 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:16.205 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:16.205 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:16.205 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:16.205 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.205 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.205 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.205 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:16.205 "name": "Existed_Raid", 00:10:16.205 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:16.205 "strip_size_kb": 64, 00:10:16.205 "state": "configuring", 00:10:16.205 "raid_level": "concat", 00:10:16.205 "superblock": false, 00:10:16.205 "num_base_bdevs": 4, 00:10:16.205 "num_base_bdevs_discovered": 3, 00:10:16.205 "num_base_bdevs_operational": 4, 00:10:16.205 "base_bdevs_list": [ 00:10:16.205 { 00:10:16.205 "name": null, 00:10:16.205 "uuid": "3172f577-9eb3-4141-9cee-19ef20e5e166", 00:10:16.205 "is_configured": false, 00:10:16.205 "data_offset": 0, 00:10:16.205 "data_size": 65536 00:10:16.205 }, 00:10:16.206 { 00:10:16.206 "name": "BaseBdev2", 00:10:16.206 "uuid": "2d09c754-4550-49ac-ab3d-4738e4e9a3cc", 00:10:16.206 "is_configured": true, 00:10:16.206 "data_offset": 0, 00:10:16.206 "data_size": 65536 00:10:16.206 }, 00:10:16.206 { 00:10:16.206 "name": "BaseBdev3", 00:10:16.206 "uuid": "dff4c5e3-ab11-48ca-974d-fa6ed16fedb6", 00:10:16.206 "is_configured": true, 00:10:16.206 "data_offset": 0, 00:10:16.206 "data_size": 65536 00:10:16.206 }, 00:10:16.206 { 00:10:16.206 "name": "BaseBdev4", 00:10:16.206 "uuid": "8f8d6298-3fc1-4c2f-929f-c741134176c9", 00:10:16.206 "is_configured": true, 00:10:16.206 "data_offset": 0, 00:10:16.206 "data_size": 65536 00:10:16.206 } 00:10:16.206 ] 00:10:16.206 }' 00:10:16.206 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:16.206 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.773 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:16.773 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.773 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.773 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:16.773 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.773 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:16.773 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:16.773 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.773 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.773 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:16.773 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.773 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 3172f577-9eb3-4141-9cee-19ef20e5e166 00:10:16.773 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.773 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.773 [2024-11-27 15:14:44.683866] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:16.773 [2024-11-27 15:14:44.683926] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:10:16.773 [2024-11-27 15:14:44.683935] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:10:16.773 [2024-11-27 15:14:44.684206] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:10:16.773 [2024-11-27 15:14:44.684327] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:10:16.773 [2024-11-27 15:14:44.684344] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006d00 00:10:16.773 [2024-11-27 15:14:44.684510] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:16.773 NewBaseBdev 00:10:16.773 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.773 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:16.774 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:10:16.774 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:16.774 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:16.774 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:16.774 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:16.774 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:16.774 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.774 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.774 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.774 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:16.774 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.774 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.774 [ 00:10:16.774 { 00:10:16.774 "name": "NewBaseBdev", 00:10:16.774 "aliases": [ 00:10:16.774 "3172f577-9eb3-4141-9cee-19ef20e5e166" 00:10:16.774 ], 00:10:16.774 "product_name": "Malloc disk", 00:10:16.774 "block_size": 512, 00:10:16.774 "num_blocks": 65536, 00:10:16.774 "uuid": "3172f577-9eb3-4141-9cee-19ef20e5e166", 00:10:16.774 "assigned_rate_limits": { 00:10:16.774 "rw_ios_per_sec": 0, 00:10:16.774 "rw_mbytes_per_sec": 0, 00:10:16.774 "r_mbytes_per_sec": 0, 00:10:16.774 "w_mbytes_per_sec": 0 00:10:16.774 }, 00:10:16.774 "claimed": true, 00:10:16.774 "claim_type": "exclusive_write", 00:10:16.774 "zoned": false, 00:10:16.774 "supported_io_types": { 00:10:16.774 "read": true, 00:10:16.774 "write": true, 00:10:16.774 "unmap": true, 00:10:16.774 "flush": true, 00:10:16.774 "reset": true, 00:10:16.774 "nvme_admin": false, 00:10:16.774 "nvme_io": false, 00:10:16.774 "nvme_io_md": false, 00:10:16.774 "write_zeroes": true, 00:10:16.774 "zcopy": true, 00:10:16.774 "get_zone_info": false, 00:10:16.774 "zone_management": false, 00:10:16.774 "zone_append": false, 00:10:16.774 "compare": false, 00:10:16.774 "compare_and_write": false, 00:10:16.774 "abort": true, 00:10:16.774 "seek_hole": false, 00:10:16.774 "seek_data": false, 00:10:16.774 "copy": true, 00:10:16.774 "nvme_iov_md": false 00:10:16.774 }, 00:10:16.774 "memory_domains": [ 00:10:16.774 { 00:10:16.774 "dma_device_id": "system", 00:10:16.774 "dma_device_type": 1 00:10:16.774 }, 00:10:16.774 { 00:10:16.774 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:16.774 "dma_device_type": 2 00:10:16.774 } 00:10:16.774 ], 00:10:16.774 "driver_specific": {} 00:10:16.774 } 00:10:16.774 ] 00:10:16.774 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.774 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:16.774 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:10:16.774 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:16.774 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:16.774 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:16.774 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:16.774 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:16.774 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:16.774 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:16.774 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:16.774 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:16.774 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:16.774 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.774 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.774 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:16.774 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.774 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:16.774 "name": "Existed_Raid", 00:10:16.774 "uuid": "987ad23b-2b12-451b-ae70-d7b9e144d682", 00:10:16.774 "strip_size_kb": 64, 00:10:16.774 "state": "online", 00:10:16.774 "raid_level": "concat", 00:10:16.774 "superblock": false, 00:10:16.774 "num_base_bdevs": 4, 00:10:16.774 "num_base_bdevs_discovered": 4, 00:10:16.774 "num_base_bdevs_operational": 4, 00:10:16.774 "base_bdevs_list": [ 00:10:16.774 { 00:10:16.774 "name": "NewBaseBdev", 00:10:16.774 "uuid": "3172f577-9eb3-4141-9cee-19ef20e5e166", 00:10:16.774 "is_configured": true, 00:10:16.774 "data_offset": 0, 00:10:16.774 "data_size": 65536 00:10:16.774 }, 00:10:16.774 { 00:10:16.774 "name": "BaseBdev2", 00:10:16.774 "uuid": "2d09c754-4550-49ac-ab3d-4738e4e9a3cc", 00:10:16.774 "is_configured": true, 00:10:16.774 "data_offset": 0, 00:10:16.774 "data_size": 65536 00:10:16.774 }, 00:10:16.774 { 00:10:16.774 "name": "BaseBdev3", 00:10:16.774 "uuid": "dff4c5e3-ab11-48ca-974d-fa6ed16fedb6", 00:10:16.774 "is_configured": true, 00:10:16.774 "data_offset": 0, 00:10:16.774 "data_size": 65536 00:10:16.774 }, 00:10:16.774 { 00:10:16.774 "name": "BaseBdev4", 00:10:16.774 "uuid": "8f8d6298-3fc1-4c2f-929f-c741134176c9", 00:10:16.774 "is_configured": true, 00:10:16.774 "data_offset": 0, 00:10:16.774 "data_size": 65536 00:10:16.774 } 00:10:16.774 ] 00:10:16.774 }' 00:10:16.774 15:14:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:16.774 15:14:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.033 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:17.033 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:17.033 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:17.033 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:17.033 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:17.033 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:17.033 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:17.033 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:17.033 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.033 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.033 [2024-11-27 15:14:45.111478] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:17.033 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:17.293 "name": "Existed_Raid", 00:10:17.293 "aliases": [ 00:10:17.293 "987ad23b-2b12-451b-ae70-d7b9e144d682" 00:10:17.293 ], 00:10:17.293 "product_name": "Raid Volume", 00:10:17.293 "block_size": 512, 00:10:17.293 "num_blocks": 262144, 00:10:17.293 "uuid": "987ad23b-2b12-451b-ae70-d7b9e144d682", 00:10:17.293 "assigned_rate_limits": { 00:10:17.293 "rw_ios_per_sec": 0, 00:10:17.293 "rw_mbytes_per_sec": 0, 00:10:17.293 "r_mbytes_per_sec": 0, 00:10:17.293 "w_mbytes_per_sec": 0 00:10:17.293 }, 00:10:17.293 "claimed": false, 00:10:17.293 "zoned": false, 00:10:17.293 "supported_io_types": { 00:10:17.293 "read": true, 00:10:17.293 "write": true, 00:10:17.293 "unmap": true, 00:10:17.293 "flush": true, 00:10:17.293 "reset": true, 00:10:17.293 "nvme_admin": false, 00:10:17.293 "nvme_io": false, 00:10:17.293 "nvme_io_md": false, 00:10:17.293 "write_zeroes": true, 00:10:17.293 "zcopy": false, 00:10:17.293 "get_zone_info": false, 00:10:17.293 "zone_management": false, 00:10:17.293 "zone_append": false, 00:10:17.293 "compare": false, 00:10:17.293 "compare_and_write": false, 00:10:17.293 "abort": false, 00:10:17.293 "seek_hole": false, 00:10:17.293 "seek_data": false, 00:10:17.293 "copy": false, 00:10:17.293 "nvme_iov_md": false 00:10:17.293 }, 00:10:17.293 "memory_domains": [ 00:10:17.293 { 00:10:17.293 "dma_device_id": "system", 00:10:17.293 "dma_device_type": 1 00:10:17.293 }, 00:10:17.293 { 00:10:17.293 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:17.293 "dma_device_type": 2 00:10:17.293 }, 00:10:17.293 { 00:10:17.293 "dma_device_id": "system", 00:10:17.293 "dma_device_type": 1 00:10:17.293 }, 00:10:17.293 { 00:10:17.293 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:17.293 "dma_device_type": 2 00:10:17.293 }, 00:10:17.293 { 00:10:17.293 "dma_device_id": "system", 00:10:17.293 "dma_device_type": 1 00:10:17.293 }, 00:10:17.293 { 00:10:17.293 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:17.293 "dma_device_type": 2 00:10:17.293 }, 00:10:17.293 { 00:10:17.293 "dma_device_id": "system", 00:10:17.293 "dma_device_type": 1 00:10:17.293 }, 00:10:17.293 { 00:10:17.293 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:17.293 "dma_device_type": 2 00:10:17.293 } 00:10:17.293 ], 00:10:17.293 "driver_specific": { 00:10:17.293 "raid": { 00:10:17.293 "uuid": "987ad23b-2b12-451b-ae70-d7b9e144d682", 00:10:17.293 "strip_size_kb": 64, 00:10:17.293 "state": "online", 00:10:17.293 "raid_level": "concat", 00:10:17.293 "superblock": false, 00:10:17.293 "num_base_bdevs": 4, 00:10:17.293 "num_base_bdevs_discovered": 4, 00:10:17.293 "num_base_bdevs_operational": 4, 00:10:17.293 "base_bdevs_list": [ 00:10:17.293 { 00:10:17.293 "name": "NewBaseBdev", 00:10:17.293 "uuid": "3172f577-9eb3-4141-9cee-19ef20e5e166", 00:10:17.293 "is_configured": true, 00:10:17.293 "data_offset": 0, 00:10:17.293 "data_size": 65536 00:10:17.293 }, 00:10:17.293 { 00:10:17.293 "name": "BaseBdev2", 00:10:17.293 "uuid": "2d09c754-4550-49ac-ab3d-4738e4e9a3cc", 00:10:17.293 "is_configured": true, 00:10:17.293 "data_offset": 0, 00:10:17.293 "data_size": 65536 00:10:17.293 }, 00:10:17.293 { 00:10:17.293 "name": "BaseBdev3", 00:10:17.293 "uuid": "dff4c5e3-ab11-48ca-974d-fa6ed16fedb6", 00:10:17.293 "is_configured": true, 00:10:17.293 "data_offset": 0, 00:10:17.293 "data_size": 65536 00:10:17.293 }, 00:10:17.293 { 00:10:17.293 "name": "BaseBdev4", 00:10:17.293 "uuid": "8f8d6298-3fc1-4c2f-929f-c741134176c9", 00:10:17.293 "is_configured": true, 00:10:17.293 "data_offset": 0, 00:10:17.293 "data_size": 65536 00:10:17.293 } 00:10:17.293 ] 00:10:17.293 } 00:10:17.293 } 00:10:17.293 }' 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:17.293 BaseBdev2 00:10:17.293 BaseBdev3 00:10:17.293 BaseBdev4' 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:17.293 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.553 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:17.553 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:17.553 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:17.553 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.553 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.553 [2024-11-27 15:14:45.430609] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:17.553 [2024-11-27 15:14:45.430640] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:17.553 [2024-11-27 15:14:45.430709] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:17.553 [2024-11-27 15:14:45.430775] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:17.553 [2024-11-27 15:14:45.430792] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name Existed_Raid, state offline 00:10:17.553 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.553 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 82321 00:10:17.553 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 82321 ']' 00:10:17.553 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 82321 00:10:17.553 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:10:17.553 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:17.553 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 82321 00:10:17.553 killing process with pid 82321 00:10:17.553 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:17.553 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:17.553 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 82321' 00:10:17.553 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 82321 00:10:17.553 [2024-11-27 15:14:45.468628] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:17.553 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 82321 00:10:17.553 [2024-11-27 15:14:45.509528] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:10:17.812 00:10:17.812 real 0m9.172s 00:10:17.812 user 0m15.591s 00:10:17.812 sys 0m1.978s 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.812 ************************************ 00:10:17.812 END TEST raid_state_function_test 00:10:17.812 ************************************ 00:10:17.812 15:14:45 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 4 true 00:10:17.812 15:14:45 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:17.812 15:14:45 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:17.812 15:14:45 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:17.812 ************************************ 00:10:17.812 START TEST raid_state_function_test_sb 00:10:17.812 ************************************ 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 4 true 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=82970 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 82970' 00:10:17.812 Process raid pid: 82970 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 82970 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 82970 ']' 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:17.812 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:17.812 15:14:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:17.812 [2024-11-27 15:14:45.891942] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:10:17.812 [2024-11-27 15:14:45.892071] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:18.072 [2024-11-27 15:14:46.062696] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:18.072 [2024-11-27 15:14:46.089138] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:18.072 [2024-11-27 15:14:46.131885] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:18.072 [2024-11-27 15:14:46.131941] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:18.641 15:14:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:18.641 15:14:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:10:18.641 15:14:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:18.641 15:14:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.641 15:14:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:18.900 [2024-11-27 15:14:46.747029] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:18.900 [2024-11-27 15:14:46.747086] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:18.900 [2024-11-27 15:14:46.747126] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:18.900 [2024-11-27 15:14:46.747137] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:18.900 [2024-11-27 15:14:46.747144] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:18.900 [2024-11-27 15:14:46.747155] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:18.900 [2024-11-27 15:14:46.747168] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:18.900 [2024-11-27 15:14:46.747176] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:18.900 15:14:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.900 15:14:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:18.900 15:14:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:18.900 15:14:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:18.900 15:14:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:18.900 15:14:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:18.900 15:14:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:18.900 15:14:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:18.900 15:14:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:18.900 15:14:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:18.900 15:14:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:18.900 15:14:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:18.900 15:14:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:18.900 15:14:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.900 15:14:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:18.900 15:14:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.900 15:14:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:18.900 "name": "Existed_Raid", 00:10:18.900 "uuid": "ae4a4ff0-2c64-4212-8c2d-0f56b1671e81", 00:10:18.900 "strip_size_kb": 64, 00:10:18.900 "state": "configuring", 00:10:18.900 "raid_level": "concat", 00:10:18.900 "superblock": true, 00:10:18.900 "num_base_bdevs": 4, 00:10:18.900 "num_base_bdevs_discovered": 0, 00:10:18.900 "num_base_bdevs_operational": 4, 00:10:18.900 "base_bdevs_list": [ 00:10:18.900 { 00:10:18.900 "name": "BaseBdev1", 00:10:18.900 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:18.900 "is_configured": false, 00:10:18.900 "data_offset": 0, 00:10:18.900 "data_size": 0 00:10:18.900 }, 00:10:18.900 { 00:10:18.900 "name": "BaseBdev2", 00:10:18.900 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:18.900 "is_configured": false, 00:10:18.900 "data_offset": 0, 00:10:18.900 "data_size": 0 00:10:18.900 }, 00:10:18.900 { 00:10:18.900 "name": "BaseBdev3", 00:10:18.900 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:18.900 "is_configured": false, 00:10:18.900 "data_offset": 0, 00:10:18.900 "data_size": 0 00:10:18.900 }, 00:10:18.900 { 00:10:18.900 "name": "BaseBdev4", 00:10:18.900 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:18.900 "is_configured": false, 00:10:18.900 "data_offset": 0, 00:10:18.900 "data_size": 0 00:10:18.900 } 00:10:18.900 ] 00:10:18.900 }' 00:10:18.900 15:14:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:18.900 15:14:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:19.160 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:19.160 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.160 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:19.160 [2024-11-27 15:14:47.214082] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:19.160 [2024-11-27 15:14:47.214123] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:10:19.160 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.160 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:19.160 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.160 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:19.160 [2024-11-27 15:14:47.226103] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:19.160 [2024-11-27 15:14:47.226144] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:19.160 [2024-11-27 15:14:47.226153] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:19.160 [2024-11-27 15:14:47.226162] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:19.160 [2024-11-27 15:14:47.226168] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:19.160 [2024-11-27 15:14:47.226177] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:19.160 [2024-11-27 15:14:47.226183] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:19.160 [2024-11-27 15:14:47.226192] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:19.160 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.160 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:19.160 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.160 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:19.160 [2024-11-27 15:14:47.246975] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:19.160 BaseBdev1 00:10:19.160 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.160 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:10:19.160 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:19.160 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:19.160 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:19.160 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:19.160 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:19.160 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:19.160 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.160 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:19.160 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.160 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:19.160 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.160 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:19.420 [ 00:10:19.420 { 00:10:19.420 "name": "BaseBdev1", 00:10:19.420 "aliases": [ 00:10:19.420 "52350f5e-d804-475c-9306-d16cfdaa95e6" 00:10:19.420 ], 00:10:19.420 "product_name": "Malloc disk", 00:10:19.420 "block_size": 512, 00:10:19.420 "num_blocks": 65536, 00:10:19.420 "uuid": "52350f5e-d804-475c-9306-d16cfdaa95e6", 00:10:19.420 "assigned_rate_limits": { 00:10:19.420 "rw_ios_per_sec": 0, 00:10:19.420 "rw_mbytes_per_sec": 0, 00:10:19.420 "r_mbytes_per_sec": 0, 00:10:19.420 "w_mbytes_per_sec": 0 00:10:19.420 }, 00:10:19.420 "claimed": true, 00:10:19.420 "claim_type": "exclusive_write", 00:10:19.420 "zoned": false, 00:10:19.420 "supported_io_types": { 00:10:19.420 "read": true, 00:10:19.420 "write": true, 00:10:19.420 "unmap": true, 00:10:19.420 "flush": true, 00:10:19.420 "reset": true, 00:10:19.420 "nvme_admin": false, 00:10:19.420 "nvme_io": false, 00:10:19.420 "nvme_io_md": false, 00:10:19.420 "write_zeroes": true, 00:10:19.420 "zcopy": true, 00:10:19.420 "get_zone_info": false, 00:10:19.420 "zone_management": false, 00:10:19.420 "zone_append": false, 00:10:19.420 "compare": false, 00:10:19.420 "compare_and_write": false, 00:10:19.420 "abort": true, 00:10:19.420 "seek_hole": false, 00:10:19.420 "seek_data": false, 00:10:19.420 "copy": true, 00:10:19.420 "nvme_iov_md": false 00:10:19.420 }, 00:10:19.420 "memory_domains": [ 00:10:19.420 { 00:10:19.420 "dma_device_id": "system", 00:10:19.420 "dma_device_type": 1 00:10:19.420 }, 00:10:19.420 { 00:10:19.420 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:19.420 "dma_device_type": 2 00:10:19.420 } 00:10:19.420 ], 00:10:19.420 "driver_specific": {} 00:10:19.420 } 00:10:19.420 ] 00:10:19.420 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.420 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:19.420 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:19.420 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:19.420 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:19.420 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:19.420 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:19.420 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:19.420 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:19.420 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:19.420 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:19.420 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:19.420 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:19.420 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:19.420 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.420 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:19.420 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.420 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:19.420 "name": "Existed_Raid", 00:10:19.420 "uuid": "ce5c23c0-f74b-4cbd-af6d-5a5496f24b2e", 00:10:19.420 "strip_size_kb": 64, 00:10:19.420 "state": "configuring", 00:10:19.420 "raid_level": "concat", 00:10:19.420 "superblock": true, 00:10:19.420 "num_base_bdevs": 4, 00:10:19.420 "num_base_bdevs_discovered": 1, 00:10:19.420 "num_base_bdevs_operational": 4, 00:10:19.420 "base_bdevs_list": [ 00:10:19.420 { 00:10:19.420 "name": "BaseBdev1", 00:10:19.420 "uuid": "52350f5e-d804-475c-9306-d16cfdaa95e6", 00:10:19.420 "is_configured": true, 00:10:19.420 "data_offset": 2048, 00:10:19.420 "data_size": 63488 00:10:19.420 }, 00:10:19.420 { 00:10:19.420 "name": "BaseBdev2", 00:10:19.420 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:19.420 "is_configured": false, 00:10:19.420 "data_offset": 0, 00:10:19.420 "data_size": 0 00:10:19.420 }, 00:10:19.420 { 00:10:19.420 "name": "BaseBdev3", 00:10:19.420 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:19.420 "is_configured": false, 00:10:19.420 "data_offset": 0, 00:10:19.420 "data_size": 0 00:10:19.420 }, 00:10:19.420 { 00:10:19.420 "name": "BaseBdev4", 00:10:19.420 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:19.420 "is_configured": false, 00:10:19.420 "data_offset": 0, 00:10:19.420 "data_size": 0 00:10:19.420 } 00:10:19.420 ] 00:10:19.420 }' 00:10:19.420 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:19.420 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:19.680 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:19.680 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.680 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:19.680 [2024-11-27 15:14:47.698256] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:19.680 [2024-11-27 15:14:47.698325] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:10:19.680 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.680 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:19.680 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.680 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:19.680 [2024-11-27 15:14:47.710284] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:19.680 [2024-11-27 15:14:47.712183] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:19.680 [2024-11-27 15:14:47.712229] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:19.680 [2024-11-27 15:14:47.712239] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:19.680 [2024-11-27 15:14:47.712247] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:19.680 [2024-11-27 15:14:47.712254] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:19.680 [2024-11-27 15:14:47.712263] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:19.680 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.680 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:10:19.680 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:19.680 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:19.680 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:19.680 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:19.680 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:19.680 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:19.680 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:19.680 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:19.680 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:19.680 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:19.680 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:19.680 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:19.680 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:19.680 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.681 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:19.681 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.681 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:19.681 "name": "Existed_Raid", 00:10:19.681 "uuid": "0976d9fd-98d7-43fa-8e9e-362a3b7c52fb", 00:10:19.681 "strip_size_kb": 64, 00:10:19.681 "state": "configuring", 00:10:19.681 "raid_level": "concat", 00:10:19.681 "superblock": true, 00:10:19.681 "num_base_bdevs": 4, 00:10:19.681 "num_base_bdevs_discovered": 1, 00:10:19.681 "num_base_bdevs_operational": 4, 00:10:19.681 "base_bdevs_list": [ 00:10:19.681 { 00:10:19.681 "name": "BaseBdev1", 00:10:19.681 "uuid": "52350f5e-d804-475c-9306-d16cfdaa95e6", 00:10:19.681 "is_configured": true, 00:10:19.681 "data_offset": 2048, 00:10:19.681 "data_size": 63488 00:10:19.681 }, 00:10:19.681 { 00:10:19.681 "name": "BaseBdev2", 00:10:19.681 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:19.681 "is_configured": false, 00:10:19.681 "data_offset": 0, 00:10:19.681 "data_size": 0 00:10:19.681 }, 00:10:19.681 { 00:10:19.681 "name": "BaseBdev3", 00:10:19.681 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:19.681 "is_configured": false, 00:10:19.681 "data_offset": 0, 00:10:19.681 "data_size": 0 00:10:19.681 }, 00:10:19.681 { 00:10:19.681 "name": "BaseBdev4", 00:10:19.681 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:19.681 "is_configured": false, 00:10:19.681 "data_offset": 0, 00:10:19.681 "data_size": 0 00:10:19.681 } 00:10:19.681 ] 00:10:19.681 }' 00:10:19.681 15:14:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:19.681 15:14:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:20.250 [2024-11-27 15:14:48.128627] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:20.250 BaseBdev2 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:20.250 [ 00:10:20.250 { 00:10:20.250 "name": "BaseBdev2", 00:10:20.250 "aliases": [ 00:10:20.250 "df635de7-dd45-4613-983c-cd13d8aedf83" 00:10:20.250 ], 00:10:20.250 "product_name": "Malloc disk", 00:10:20.250 "block_size": 512, 00:10:20.250 "num_blocks": 65536, 00:10:20.250 "uuid": "df635de7-dd45-4613-983c-cd13d8aedf83", 00:10:20.250 "assigned_rate_limits": { 00:10:20.250 "rw_ios_per_sec": 0, 00:10:20.250 "rw_mbytes_per_sec": 0, 00:10:20.250 "r_mbytes_per_sec": 0, 00:10:20.250 "w_mbytes_per_sec": 0 00:10:20.250 }, 00:10:20.250 "claimed": true, 00:10:20.250 "claim_type": "exclusive_write", 00:10:20.250 "zoned": false, 00:10:20.250 "supported_io_types": { 00:10:20.250 "read": true, 00:10:20.250 "write": true, 00:10:20.250 "unmap": true, 00:10:20.250 "flush": true, 00:10:20.250 "reset": true, 00:10:20.250 "nvme_admin": false, 00:10:20.250 "nvme_io": false, 00:10:20.250 "nvme_io_md": false, 00:10:20.250 "write_zeroes": true, 00:10:20.250 "zcopy": true, 00:10:20.250 "get_zone_info": false, 00:10:20.250 "zone_management": false, 00:10:20.250 "zone_append": false, 00:10:20.250 "compare": false, 00:10:20.250 "compare_and_write": false, 00:10:20.250 "abort": true, 00:10:20.250 "seek_hole": false, 00:10:20.250 "seek_data": false, 00:10:20.250 "copy": true, 00:10:20.250 "nvme_iov_md": false 00:10:20.250 }, 00:10:20.250 "memory_domains": [ 00:10:20.250 { 00:10:20.250 "dma_device_id": "system", 00:10:20.250 "dma_device_type": 1 00:10:20.250 }, 00:10:20.250 { 00:10:20.250 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:20.250 "dma_device_type": 2 00:10:20.250 } 00:10:20.250 ], 00:10:20.250 "driver_specific": {} 00:10:20.250 } 00:10:20.250 ] 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:20.250 "name": "Existed_Raid", 00:10:20.250 "uuid": "0976d9fd-98d7-43fa-8e9e-362a3b7c52fb", 00:10:20.250 "strip_size_kb": 64, 00:10:20.250 "state": "configuring", 00:10:20.250 "raid_level": "concat", 00:10:20.250 "superblock": true, 00:10:20.250 "num_base_bdevs": 4, 00:10:20.250 "num_base_bdevs_discovered": 2, 00:10:20.250 "num_base_bdevs_operational": 4, 00:10:20.250 "base_bdevs_list": [ 00:10:20.250 { 00:10:20.250 "name": "BaseBdev1", 00:10:20.250 "uuid": "52350f5e-d804-475c-9306-d16cfdaa95e6", 00:10:20.250 "is_configured": true, 00:10:20.250 "data_offset": 2048, 00:10:20.250 "data_size": 63488 00:10:20.250 }, 00:10:20.250 { 00:10:20.250 "name": "BaseBdev2", 00:10:20.250 "uuid": "df635de7-dd45-4613-983c-cd13d8aedf83", 00:10:20.250 "is_configured": true, 00:10:20.250 "data_offset": 2048, 00:10:20.250 "data_size": 63488 00:10:20.250 }, 00:10:20.250 { 00:10:20.250 "name": "BaseBdev3", 00:10:20.250 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:20.250 "is_configured": false, 00:10:20.250 "data_offset": 0, 00:10:20.250 "data_size": 0 00:10:20.250 }, 00:10:20.250 { 00:10:20.250 "name": "BaseBdev4", 00:10:20.250 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:20.250 "is_configured": false, 00:10:20.250 "data_offset": 0, 00:10:20.250 "data_size": 0 00:10:20.250 } 00:10:20.250 ] 00:10:20.250 }' 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:20.250 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:20.511 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:20.511 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.511 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:20.511 [2024-11-27 15:14:48.588936] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:20.511 BaseBdev3 00:10:20.511 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.511 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:10:20.511 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:20.511 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:20.511 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:20.511 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:20.511 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:20.511 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:20.511 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.511 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:20.511 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.511 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:20.511 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.511 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:20.511 [ 00:10:20.511 { 00:10:20.511 "name": "BaseBdev3", 00:10:20.511 "aliases": [ 00:10:20.511 "28ae04b1-244d-471e-a39e-2bd577d2e1d1" 00:10:20.511 ], 00:10:20.511 "product_name": "Malloc disk", 00:10:20.511 "block_size": 512, 00:10:20.511 "num_blocks": 65536, 00:10:20.511 "uuid": "28ae04b1-244d-471e-a39e-2bd577d2e1d1", 00:10:20.511 "assigned_rate_limits": { 00:10:20.511 "rw_ios_per_sec": 0, 00:10:20.511 "rw_mbytes_per_sec": 0, 00:10:20.511 "r_mbytes_per_sec": 0, 00:10:20.511 "w_mbytes_per_sec": 0 00:10:20.511 }, 00:10:20.511 "claimed": true, 00:10:20.511 "claim_type": "exclusive_write", 00:10:20.511 "zoned": false, 00:10:20.511 "supported_io_types": { 00:10:20.511 "read": true, 00:10:20.511 "write": true, 00:10:20.511 "unmap": true, 00:10:20.511 "flush": true, 00:10:20.511 "reset": true, 00:10:20.511 "nvme_admin": false, 00:10:20.511 "nvme_io": false, 00:10:20.511 "nvme_io_md": false, 00:10:20.511 "write_zeroes": true, 00:10:20.511 "zcopy": true, 00:10:20.511 "get_zone_info": false, 00:10:20.771 "zone_management": false, 00:10:20.771 "zone_append": false, 00:10:20.771 "compare": false, 00:10:20.771 "compare_and_write": false, 00:10:20.771 "abort": true, 00:10:20.771 "seek_hole": false, 00:10:20.771 "seek_data": false, 00:10:20.771 "copy": true, 00:10:20.771 "nvme_iov_md": false 00:10:20.771 }, 00:10:20.771 "memory_domains": [ 00:10:20.771 { 00:10:20.771 "dma_device_id": "system", 00:10:20.771 "dma_device_type": 1 00:10:20.771 }, 00:10:20.771 { 00:10:20.771 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:20.771 "dma_device_type": 2 00:10:20.771 } 00:10:20.771 ], 00:10:20.771 "driver_specific": {} 00:10:20.771 } 00:10:20.771 ] 00:10:20.771 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.771 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:20.771 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:20.771 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:20.771 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:20.771 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:20.771 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:20.771 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:20.771 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:20.771 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:20.771 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:20.771 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:20.772 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:20.772 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:20.772 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:20.772 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.772 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:20.772 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:20.772 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.772 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:20.772 "name": "Existed_Raid", 00:10:20.772 "uuid": "0976d9fd-98d7-43fa-8e9e-362a3b7c52fb", 00:10:20.772 "strip_size_kb": 64, 00:10:20.772 "state": "configuring", 00:10:20.772 "raid_level": "concat", 00:10:20.772 "superblock": true, 00:10:20.772 "num_base_bdevs": 4, 00:10:20.772 "num_base_bdevs_discovered": 3, 00:10:20.772 "num_base_bdevs_operational": 4, 00:10:20.772 "base_bdevs_list": [ 00:10:20.772 { 00:10:20.772 "name": "BaseBdev1", 00:10:20.772 "uuid": "52350f5e-d804-475c-9306-d16cfdaa95e6", 00:10:20.772 "is_configured": true, 00:10:20.772 "data_offset": 2048, 00:10:20.772 "data_size": 63488 00:10:20.772 }, 00:10:20.772 { 00:10:20.772 "name": "BaseBdev2", 00:10:20.772 "uuid": "df635de7-dd45-4613-983c-cd13d8aedf83", 00:10:20.772 "is_configured": true, 00:10:20.772 "data_offset": 2048, 00:10:20.772 "data_size": 63488 00:10:20.772 }, 00:10:20.772 { 00:10:20.772 "name": "BaseBdev3", 00:10:20.772 "uuid": "28ae04b1-244d-471e-a39e-2bd577d2e1d1", 00:10:20.772 "is_configured": true, 00:10:20.772 "data_offset": 2048, 00:10:20.772 "data_size": 63488 00:10:20.772 }, 00:10:20.772 { 00:10:20.772 "name": "BaseBdev4", 00:10:20.772 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:20.772 "is_configured": false, 00:10:20.772 "data_offset": 0, 00:10:20.772 "data_size": 0 00:10:20.772 } 00:10:20.772 ] 00:10:20.772 }' 00:10:20.772 15:14:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:20.772 15:14:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:21.032 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:21.032 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.032 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:21.032 [2024-11-27 15:14:49.071148] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:21.032 [2024-11-27 15:14:49.071369] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:10:21.032 [2024-11-27 15:14:49.071384] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:21.032 [2024-11-27 15:14:49.071659] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:10:21.032 BaseBdev4 00:10:21.032 [2024-11-27 15:14:49.071793] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:10:21.032 [2024-11-27 15:14:49.071807] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:10:21.032 [2024-11-27 15:14:49.071925] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:21.032 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.032 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:10:21.032 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:10:21.032 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:21.032 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:21.032 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:21.032 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:21.032 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:21.032 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.032 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:21.032 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.032 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:21.032 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.032 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:21.032 [ 00:10:21.032 { 00:10:21.032 "name": "BaseBdev4", 00:10:21.032 "aliases": [ 00:10:21.032 "3b4134f6-9ab5-415d-af54-29d033c27098" 00:10:21.032 ], 00:10:21.032 "product_name": "Malloc disk", 00:10:21.032 "block_size": 512, 00:10:21.032 "num_blocks": 65536, 00:10:21.032 "uuid": "3b4134f6-9ab5-415d-af54-29d033c27098", 00:10:21.032 "assigned_rate_limits": { 00:10:21.032 "rw_ios_per_sec": 0, 00:10:21.032 "rw_mbytes_per_sec": 0, 00:10:21.032 "r_mbytes_per_sec": 0, 00:10:21.032 "w_mbytes_per_sec": 0 00:10:21.032 }, 00:10:21.032 "claimed": true, 00:10:21.032 "claim_type": "exclusive_write", 00:10:21.032 "zoned": false, 00:10:21.032 "supported_io_types": { 00:10:21.032 "read": true, 00:10:21.032 "write": true, 00:10:21.032 "unmap": true, 00:10:21.032 "flush": true, 00:10:21.032 "reset": true, 00:10:21.032 "nvme_admin": false, 00:10:21.032 "nvme_io": false, 00:10:21.032 "nvme_io_md": false, 00:10:21.032 "write_zeroes": true, 00:10:21.032 "zcopy": true, 00:10:21.032 "get_zone_info": false, 00:10:21.032 "zone_management": false, 00:10:21.032 "zone_append": false, 00:10:21.032 "compare": false, 00:10:21.032 "compare_and_write": false, 00:10:21.032 "abort": true, 00:10:21.032 "seek_hole": false, 00:10:21.032 "seek_data": false, 00:10:21.032 "copy": true, 00:10:21.032 "nvme_iov_md": false 00:10:21.032 }, 00:10:21.032 "memory_domains": [ 00:10:21.032 { 00:10:21.032 "dma_device_id": "system", 00:10:21.032 "dma_device_type": 1 00:10:21.032 }, 00:10:21.032 { 00:10:21.032 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:21.032 "dma_device_type": 2 00:10:21.032 } 00:10:21.032 ], 00:10:21.032 "driver_specific": {} 00:10:21.032 } 00:10:21.032 ] 00:10:21.032 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.032 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:21.033 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:21.033 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:21.033 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:10:21.033 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:21.033 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:21.033 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:21.033 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:21.033 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:21.033 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:21.033 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:21.033 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:21.033 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:21.033 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:21.033 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:21.033 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.033 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:21.033 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.292 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:21.292 "name": "Existed_Raid", 00:10:21.292 "uuid": "0976d9fd-98d7-43fa-8e9e-362a3b7c52fb", 00:10:21.292 "strip_size_kb": 64, 00:10:21.292 "state": "online", 00:10:21.292 "raid_level": "concat", 00:10:21.292 "superblock": true, 00:10:21.292 "num_base_bdevs": 4, 00:10:21.292 "num_base_bdevs_discovered": 4, 00:10:21.292 "num_base_bdevs_operational": 4, 00:10:21.292 "base_bdevs_list": [ 00:10:21.292 { 00:10:21.292 "name": "BaseBdev1", 00:10:21.292 "uuid": "52350f5e-d804-475c-9306-d16cfdaa95e6", 00:10:21.292 "is_configured": true, 00:10:21.292 "data_offset": 2048, 00:10:21.292 "data_size": 63488 00:10:21.292 }, 00:10:21.292 { 00:10:21.292 "name": "BaseBdev2", 00:10:21.292 "uuid": "df635de7-dd45-4613-983c-cd13d8aedf83", 00:10:21.292 "is_configured": true, 00:10:21.292 "data_offset": 2048, 00:10:21.292 "data_size": 63488 00:10:21.292 }, 00:10:21.292 { 00:10:21.292 "name": "BaseBdev3", 00:10:21.292 "uuid": "28ae04b1-244d-471e-a39e-2bd577d2e1d1", 00:10:21.292 "is_configured": true, 00:10:21.292 "data_offset": 2048, 00:10:21.292 "data_size": 63488 00:10:21.292 }, 00:10:21.292 { 00:10:21.292 "name": "BaseBdev4", 00:10:21.292 "uuid": "3b4134f6-9ab5-415d-af54-29d033c27098", 00:10:21.292 "is_configured": true, 00:10:21.292 "data_offset": 2048, 00:10:21.292 "data_size": 63488 00:10:21.292 } 00:10:21.292 ] 00:10:21.292 }' 00:10:21.292 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:21.292 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:21.552 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:21.552 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:21.552 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:21.552 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:21.552 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:21.552 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:21.552 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:21.552 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:21.552 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.552 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:21.552 [2024-11-27 15:14:49.566705] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:21.552 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.552 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:21.552 "name": "Existed_Raid", 00:10:21.552 "aliases": [ 00:10:21.552 "0976d9fd-98d7-43fa-8e9e-362a3b7c52fb" 00:10:21.552 ], 00:10:21.552 "product_name": "Raid Volume", 00:10:21.552 "block_size": 512, 00:10:21.552 "num_blocks": 253952, 00:10:21.552 "uuid": "0976d9fd-98d7-43fa-8e9e-362a3b7c52fb", 00:10:21.552 "assigned_rate_limits": { 00:10:21.552 "rw_ios_per_sec": 0, 00:10:21.552 "rw_mbytes_per_sec": 0, 00:10:21.552 "r_mbytes_per_sec": 0, 00:10:21.552 "w_mbytes_per_sec": 0 00:10:21.552 }, 00:10:21.552 "claimed": false, 00:10:21.552 "zoned": false, 00:10:21.552 "supported_io_types": { 00:10:21.552 "read": true, 00:10:21.552 "write": true, 00:10:21.552 "unmap": true, 00:10:21.552 "flush": true, 00:10:21.552 "reset": true, 00:10:21.552 "nvme_admin": false, 00:10:21.552 "nvme_io": false, 00:10:21.552 "nvme_io_md": false, 00:10:21.552 "write_zeroes": true, 00:10:21.552 "zcopy": false, 00:10:21.552 "get_zone_info": false, 00:10:21.552 "zone_management": false, 00:10:21.552 "zone_append": false, 00:10:21.552 "compare": false, 00:10:21.552 "compare_and_write": false, 00:10:21.552 "abort": false, 00:10:21.552 "seek_hole": false, 00:10:21.552 "seek_data": false, 00:10:21.552 "copy": false, 00:10:21.552 "nvme_iov_md": false 00:10:21.552 }, 00:10:21.552 "memory_domains": [ 00:10:21.552 { 00:10:21.552 "dma_device_id": "system", 00:10:21.552 "dma_device_type": 1 00:10:21.552 }, 00:10:21.552 { 00:10:21.552 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:21.552 "dma_device_type": 2 00:10:21.552 }, 00:10:21.552 { 00:10:21.552 "dma_device_id": "system", 00:10:21.552 "dma_device_type": 1 00:10:21.552 }, 00:10:21.552 { 00:10:21.552 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:21.552 "dma_device_type": 2 00:10:21.552 }, 00:10:21.552 { 00:10:21.552 "dma_device_id": "system", 00:10:21.552 "dma_device_type": 1 00:10:21.552 }, 00:10:21.552 { 00:10:21.552 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:21.552 "dma_device_type": 2 00:10:21.552 }, 00:10:21.552 { 00:10:21.552 "dma_device_id": "system", 00:10:21.552 "dma_device_type": 1 00:10:21.552 }, 00:10:21.552 { 00:10:21.552 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:21.552 "dma_device_type": 2 00:10:21.552 } 00:10:21.552 ], 00:10:21.552 "driver_specific": { 00:10:21.552 "raid": { 00:10:21.552 "uuid": "0976d9fd-98d7-43fa-8e9e-362a3b7c52fb", 00:10:21.552 "strip_size_kb": 64, 00:10:21.552 "state": "online", 00:10:21.552 "raid_level": "concat", 00:10:21.552 "superblock": true, 00:10:21.552 "num_base_bdevs": 4, 00:10:21.552 "num_base_bdevs_discovered": 4, 00:10:21.552 "num_base_bdevs_operational": 4, 00:10:21.552 "base_bdevs_list": [ 00:10:21.552 { 00:10:21.552 "name": "BaseBdev1", 00:10:21.552 "uuid": "52350f5e-d804-475c-9306-d16cfdaa95e6", 00:10:21.552 "is_configured": true, 00:10:21.552 "data_offset": 2048, 00:10:21.552 "data_size": 63488 00:10:21.552 }, 00:10:21.552 { 00:10:21.552 "name": "BaseBdev2", 00:10:21.552 "uuid": "df635de7-dd45-4613-983c-cd13d8aedf83", 00:10:21.552 "is_configured": true, 00:10:21.552 "data_offset": 2048, 00:10:21.552 "data_size": 63488 00:10:21.552 }, 00:10:21.552 { 00:10:21.552 "name": "BaseBdev3", 00:10:21.552 "uuid": "28ae04b1-244d-471e-a39e-2bd577d2e1d1", 00:10:21.552 "is_configured": true, 00:10:21.552 "data_offset": 2048, 00:10:21.552 "data_size": 63488 00:10:21.552 }, 00:10:21.552 { 00:10:21.552 "name": "BaseBdev4", 00:10:21.552 "uuid": "3b4134f6-9ab5-415d-af54-29d033c27098", 00:10:21.552 "is_configured": true, 00:10:21.552 "data_offset": 2048, 00:10:21.552 "data_size": 63488 00:10:21.552 } 00:10:21.552 ] 00:10:21.552 } 00:10:21.552 } 00:10:21.552 }' 00:10:21.552 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:21.552 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:21.552 BaseBdev2 00:10:21.552 BaseBdev3 00:10:21.552 BaseBdev4' 00:10:21.552 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:21.812 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:21.812 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:21.812 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:21.812 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.812 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:21.812 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:21.812 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.812 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:21.813 [2024-11-27 15:14:49.866002] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:21.813 [2024-11-27 15:14:49.866042] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:21.813 [2024-11-27 15:14:49.866098] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 3 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:21.813 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.072 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:22.072 "name": "Existed_Raid", 00:10:22.072 "uuid": "0976d9fd-98d7-43fa-8e9e-362a3b7c52fb", 00:10:22.072 "strip_size_kb": 64, 00:10:22.072 "state": "offline", 00:10:22.072 "raid_level": "concat", 00:10:22.072 "superblock": true, 00:10:22.072 "num_base_bdevs": 4, 00:10:22.072 "num_base_bdevs_discovered": 3, 00:10:22.072 "num_base_bdevs_operational": 3, 00:10:22.072 "base_bdevs_list": [ 00:10:22.072 { 00:10:22.072 "name": null, 00:10:22.072 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:22.072 "is_configured": false, 00:10:22.072 "data_offset": 0, 00:10:22.072 "data_size": 63488 00:10:22.072 }, 00:10:22.072 { 00:10:22.072 "name": "BaseBdev2", 00:10:22.072 "uuid": "df635de7-dd45-4613-983c-cd13d8aedf83", 00:10:22.072 "is_configured": true, 00:10:22.072 "data_offset": 2048, 00:10:22.072 "data_size": 63488 00:10:22.072 }, 00:10:22.072 { 00:10:22.072 "name": "BaseBdev3", 00:10:22.072 "uuid": "28ae04b1-244d-471e-a39e-2bd577d2e1d1", 00:10:22.072 "is_configured": true, 00:10:22.072 "data_offset": 2048, 00:10:22.072 "data_size": 63488 00:10:22.072 }, 00:10:22.072 { 00:10:22.072 "name": "BaseBdev4", 00:10:22.072 "uuid": "3b4134f6-9ab5-415d-af54-29d033c27098", 00:10:22.072 "is_configured": true, 00:10:22.072 "data_offset": 2048, 00:10:22.072 "data_size": 63488 00:10:22.072 } 00:10:22.072 ] 00:10:22.072 }' 00:10:22.072 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:22.072 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.332 [2024-11-27 15:14:50.304616] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.332 [2024-11-27 15:14:50.375603] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:22.332 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.592 [2024-11-27 15:14:50.446616] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:10:22.592 [2024-11-27 15:14:50.446667] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.592 BaseBdev2 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.592 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.592 [ 00:10:22.592 { 00:10:22.592 "name": "BaseBdev2", 00:10:22.592 "aliases": [ 00:10:22.592 "ddb0aa0b-7ed5-4ae2-b6a7-cb713fa28ae8" 00:10:22.592 ], 00:10:22.593 "product_name": "Malloc disk", 00:10:22.593 "block_size": 512, 00:10:22.593 "num_blocks": 65536, 00:10:22.593 "uuid": "ddb0aa0b-7ed5-4ae2-b6a7-cb713fa28ae8", 00:10:22.593 "assigned_rate_limits": { 00:10:22.593 "rw_ios_per_sec": 0, 00:10:22.593 "rw_mbytes_per_sec": 0, 00:10:22.593 "r_mbytes_per_sec": 0, 00:10:22.593 "w_mbytes_per_sec": 0 00:10:22.593 }, 00:10:22.593 "claimed": false, 00:10:22.593 "zoned": false, 00:10:22.593 "supported_io_types": { 00:10:22.593 "read": true, 00:10:22.593 "write": true, 00:10:22.593 "unmap": true, 00:10:22.593 "flush": true, 00:10:22.593 "reset": true, 00:10:22.593 "nvme_admin": false, 00:10:22.593 "nvme_io": false, 00:10:22.593 "nvme_io_md": false, 00:10:22.593 "write_zeroes": true, 00:10:22.593 "zcopy": true, 00:10:22.593 "get_zone_info": false, 00:10:22.593 "zone_management": false, 00:10:22.593 "zone_append": false, 00:10:22.593 "compare": false, 00:10:22.593 "compare_and_write": false, 00:10:22.593 "abort": true, 00:10:22.593 "seek_hole": false, 00:10:22.593 "seek_data": false, 00:10:22.593 "copy": true, 00:10:22.593 "nvme_iov_md": false 00:10:22.593 }, 00:10:22.593 "memory_domains": [ 00:10:22.593 { 00:10:22.593 "dma_device_id": "system", 00:10:22.593 "dma_device_type": 1 00:10:22.593 }, 00:10:22.593 { 00:10:22.593 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:22.593 "dma_device_type": 2 00:10:22.593 } 00:10:22.593 ], 00:10:22.593 "driver_specific": {} 00:10:22.593 } 00:10:22.593 ] 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.593 BaseBdev3 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.593 [ 00:10:22.593 { 00:10:22.593 "name": "BaseBdev3", 00:10:22.593 "aliases": [ 00:10:22.593 "82d2514e-25e5-4f74-8790-f575e2ca382a" 00:10:22.593 ], 00:10:22.593 "product_name": "Malloc disk", 00:10:22.593 "block_size": 512, 00:10:22.593 "num_blocks": 65536, 00:10:22.593 "uuid": "82d2514e-25e5-4f74-8790-f575e2ca382a", 00:10:22.593 "assigned_rate_limits": { 00:10:22.593 "rw_ios_per_sec": 0, 00:10:22.593 "rw_mbytes_per_sec": 0, 00:10:22.593 "r_mbytes_per_sec": 0, 00:10:22.593 "w_mbytes_per_sec": 0 00:10:22.593 }, 00:10:22.593 "claimed": false, 00:10:22.593 "zoned": false, 00:10:22.593 "supported_io_types": { 00:10:22.593 "read": true, 00:10:22.593 "write": true, 00:10:22.593 "unmap": true, 00:10:22.593 "flush": true, 00:10:22.593 "reset": true, 00:10:22.593 "nvme_admin": false, 00:10:22.593 "nvme_io": false, 00:10:22.593 "nvme_io_md": false, 00:10:22.593 "write_zeroes": true, 00:10:22.593 "zcopy": true, 00:10:22.593 "get_zone_info": false, 00:10:22.593 "zone_management": false, 00:10:22.593 "zone_append": false, 00:10:22.593 "compare": false, 00:10:22.593 "compare_and_write": false, 00:10:22.593 "abort": true, 00:10:22.593 "seek_hole": false, 00:10:22.593 "seek_data": false, 00:10:22.593 "copy": true, 00:10:22.593 "nvme_iov_md": false 00:10:22.593 }, 00:10:22.593 "memory_domains": [ 00:10:22.593 { 00:10:22.593 "dma_device_id": "system", 00:10:22.593 "dma_device_type": 1 00:10:22.593 }, 00:10:22.593 { 00:10:22.593 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:22.593 "dma_device_type": 2 00:10:22.593 } 00:10:22.593 ], 00:10:22.593 "driver_specific": {} 00:10:22.593 } 00:10:22.593 ] 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.593 BaseBdev4 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.593 [ 00:10:22.593 { 00:10:22.593 "name": "BaseBdev4", 00:10:22.593 "aliases": [ 00:10:22.593 "b1d9ac8a-6359-4630-98b7-374f84ba7728" 00:10:22.593 ], 00:10:22.593 "product_name": "Malloc disk", 00:10:22.593 "block_size": 512, 00:10:22.593 "num_blocks": 65536, 00:10:22.593 "uuid": "b1d9ac8a-6359-4630-98b7-374f84ba7728", 00:10:22.593 "assigned_rate_limits": { 00:10:22.593 "rw_ios_per_sec": 0, 00:10:22.593 "rw_mbytes_per_sec": 0, 00:10:22.593 "r_mbytes_per_sec": 0, 00:10:22.593 "w_mbytes_per_sec": 0 00:10:22.593 }, 00:10:22.593 "claimed": false, 00:10:22.593 "zoned": false, 00:10:22.593 "supported_io_types": { 00:10:22.593 "read": true, 00:10:22.593 "write": true, 00:10:22.593 "unmap": true, 00:10:22.593 "flush": true, 00:10:22.593 "reset": true, 00:10:22.593 "nvme_admin": false, 00:10:22.593 "nvme_io": false, 00:10:22.593 "nvme_io_md": false, 00:10:22.593 "write_zeroes": true, 00:10:22.593 "zcopy": true, 00:10:22.593 "get_zone_info": false, 00:10:22.593 "zone_management": false, 00:10:22.593 "zone_append": false, 00:10:22.593 "compare": false, 00:10:22.593 "compare_and_write": false, 00:10:22.593 "abort": true, 00:10:22.593 "seek_hole": false, 00:10:22.593 "seek_data": false, 00:10:22.593 "copy": true, 00:10:22.593 "nvme_iov_md": false 00:10:22.593 }, 00:10:22.593 "memory_domains": [ 00:10:22.593 { 00:10:22.593 "dma_device_id": "system", 00:10:22.593 "dma_device_type": 1 00:10:22.593 }, 00:10:22.593 { 00:10:22.593 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:22.593 "dma_device_type": 2 00:10:22.593 } 00:10:22.593 ], 00:10:22.593 "driver_specific": {} 00:10:22.593 } 00:10:22.593 ] 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.593 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.593 [2024-11-27 15:14:50.667309] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:22.593 [2024-11-27 15:14:50.667353] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:22.594 [2024-11-27 15:14:50.667374] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:22.594 [2024-11-27 15:14:50.669174] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:22.594 [2024-11-27 15:14:50.669225] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:22.594 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.594 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:22.594 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:22.594 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:22.594 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:22.594 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:22.594 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:22.594 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:22.594 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:22.594 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:22.594 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:22.594 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:22.594 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.594 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.594 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:22.854 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.854 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:22.854 "name": "Existed_Raid", 00:10:22.854 "uuid": "27c11bfe-147c-47e8-82a2-0db8e51ba085", 00:10:22.854 "strip_size_kb": 64, 00:10:22.854 "state": "configuring", 00:10:22.854 "raid_level": "concat", 00:10:22.854 "superblock": true, 00:10:22.854 "num_base_bdevs": 4, 00:10:22.854 "num_base_bdevs_discovered": 3, 00:10:22.854 "num_base_bdevs_operational": 4, 00:10:22.854 "base_bdevs_list": [ 00:10:22.854 { 00:10:22.854 "name": "BaseBdev1", 00:10:22.854 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:22.854 "is_configured": false, 00:10:22.854 "data_offset": 0, 00:10:22.854 "data_size": 0 00:10:22.854 }, 00:10:22.854 { 00:10:22.854 "name": "BaseBdev2", 00:10:22.854 "uuid": "ddb0aa0b-7ed5-4ae2-b6a7-cb713fa28ae8", 00:10:22.854 "is_configured": true, 00:10:22.854 "data_offset": 2048, 00:10:22.854 "data_size": 63488 00:10:22.854 }, 00:10:22.854 { 00:10:22.854 "name": "BaseBdev3", 00:10:22.854 "uuid": "82d2514e-25e5-4f74-8790-f575e2ca382a", 00:10:22.854 "is_configured": true, 00:10:22.854 "data_offset": 2048, 00:10:22.854 "data_size": 63488 00:10:22.854 }, 00:10:22.854 { 00:10:22.854 "name": "BaseBdev4", 00:10:22.854 "uuid": "b1d9ac8a-6359-4630-98b7-374f84ba7728", 00:10:22.854 "is_configured": true, 00:10:22.854 "data_offset": 2048, 00:10:22.854 "data_size": 63488 00:10:22.854 } 00:10:22.854 ] 00:10:22.854 }' 00:10:22.854 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:22.854 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:23.114 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:23.114 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:23.114 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:23.114 [2024-11-27 15:14:51.074646] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:23.114 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:23.114 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:23.114 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:23.114 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:23.114 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:23.114 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:23.114 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:23.114 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:23.114 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:23.114 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:23.114 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:23.114 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:23.114 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:23.114 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:23.114 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:23.114 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:23.114 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:23.114 "name": "Existed_Raid", 00:10:23.114 "uuid": "27c11bfe-147c-47e8-82a2-0db8e51ba085", 00:10:23.114 "strip_size_kb": 64, 00:10:23.115 "state": "configuring", 00:10:23.115 "raid_level": "concat", 00:10:23.115 "superblock": true, 00:10:23.115 "num_base_bdevs": 4, 00:10:23.115 "num_base_bdevs_discovered": 2, 00:10:23.115 "num_base_bdevs_operational": 4, 00:10:23.115 "base_bdevs_list": [ 00:10:23.115 { 00:10:23.115 "name": "BaseBdev1", 00:10:23.115 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:23.115 "is_configured": false, 00:10:23.115 "data_offset": 0, 00:10:23.115 "data_size": 0 00:10:23.115 }, 00:10:23.115 { 00:10:23.115 "name": null, 00:10:23.115 "uuid": "ddb0aa0b-7ed5-4ae2-b6a7-cb713fa28ae8", 00:10:23.115 "is_configured": false, 00:10:23.115 "data_offset": 0, 00:10:23.115 "data_size": 63488 00:10:23.115 }, 00:10:23.115 { 00:10:23.115 "name": "BaseBdev3", 00:10:23.115 "uuid": "82d2514e-25e5-4f74-8790-f575e2ca382a", 00:10:23.115 "is_configured": true, 00:10:23.115 "data_offset": 2048, 00:10:23.115 "data_size": 63488 00:10:23.115 }, 00:10:23.115 { 00:10:23.115 "name": "BaseBdev4", 00:10:23.115 "uuid": "b1d9ac8a-6359-4630-98b7-374f84ba7728", 00:10:23.115 "is_configured": true, 00:10:23.115 "data_offset": 2048, 00:10:23.115 "data_size": 63488 00:10:23.115 } 00:10:23.115 ] 00:10:23.115 }' 00:10:23.115 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:23.115 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:23.684 [2024-11-27 15:14:51.512806] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:23.684 BaseBdev1 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:23.684 [ 00:10:23.684 { 00:10:23.684 "name": "BaseBdev1", 00:10:23.684 "aliases": [ 00:10:23.684 "2b95f3ea-5e40-4960-91d3-0105f9af776c" 00:10:23.684 ], 00:10:23.684 "product_name": "Malloc disk", 00:10:23.684 "block_size": 512, 00:10:23.684 "num_blocks": 65536, 00:10:23.684 "uuid": "2b95f3ea-5e40-4960-91d3-0105f9af776c", 00:10:23.684 "assigned_rate_limits": { 00:10:23.684 "rw_ios_per_sec": 0, 00:10:23.684 "rw_mbytes_per_sec": 0, 00:10:23.684 "r_mbytes_per_sec": 0, 00:10:23.684 "w_mbytes_per_sec": 0 00:10:23.684 }, 00:10:23.684 "claimed": true, 00:10:23.684 "claim_type": "exclusive_write", 00:10:23.684 "zoned": false, 00:10:23.684 "supported_io_types": { 00:10:23.684 "read": true, 00:10:23.684 "write": true, 00:10:23.684 "unmap": true, 00:10:23.684 "flush": true, 00:10:23.684 "reset": true, 00:10:23.684 "nvme_admin": false, 00:10:23.684 "nvme_io": false, 00:10:23.684 "nvme_io_md": false, 00:10:23.684 "write_zeroes": true, 00:10:23.684 "zcopy": true, 00:10:23.684 "get_zone_info": false, 00:10:23.684 "zone_management": false, 00:10:23.684 "zone_append": false, 00:10:23.684 "compare": false, 00:10:23.684 "compare_and_write": false, 00:10:23.684 "abort": true, 00:10:23.684 "seek_hole": false, 00:10:23.684 "seek_data": false, 00:10:23.684 "copy": true, 00:10:23.684 "nvme_iov_md": false 00:10:23.684 }, 00:10:23.684 "memory_domains": [ 00:10:23.684 { 00:10:23.684 "dma_device_id": "system", 00:10:23.684 "dma_device_type": 1 00:10:23.684 }, 00:10:23.684 { 00:10:23.684 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:23.684 "dma_device_type": 2 00:10:23.684 } 00:10:23.684 ], 00:10:23.684 "driver_specific": {} 00:10:23.684 } 00:10:23.684 ] 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:23.684 "name": "Existed_Raid", 00:10:23.684 "uuid": "27c11bfe-147c-47e8-82a2-0db8e51ba085", 00:10:23.684 "strip_size_kb": 64, 00:10:23.684 "state": "configuring", 00:10:23.684 "raid_level": "concat", 00:10:23.684 "superblock": true, 00:10:23.684 "num_base_bdevs": 4, 00:10:23.684 "num_base_bdevs_discovered": 3, 00:10:23.684 "num_base_bdevs_operational": 4, 00:10:23.684 "base_bdevs_list": [ 00:10:23.684 { 00:10:23.684 "name": "BaseBdev1", 00:10:23.684 "uuid": "2b95f3ea-5e40-4960-91d3-0105f9af776c", 00:10:23.684 "is_configured": true, 00:10:23.684 "data_offset": 2048, 00:10:23.684 "data_size": 63488 00:10:23.684 }, 00:10:23.684 { 00:10:23.684 "name": null, 00:10:23.684 "uuid": "ddb0aa0b-7ed5-4ae2-b6a7-cb713fa28ae8", 00:10:23.684 "is_configured": false, 00:10:23.684 "data_offset": 0, 00:10:23.684 "data_size": 63488 00:10:23.684 }, 00:10:23.684 { 00:10:23.684 "name": "BaseBdev3", 00:10:23.684 "uuid": "82d2514e-25e5-4f74-8790-f575e2ca382a", 00:10:23.684 "is_configured": true, 00:10:23.684 "data_offset": 2048, 00:10:23.684 "data_size": 63488 00:10:23.684 }, 00:10:23.684 { 00:10:23.684 "name": "BaseBdev4", 00:10:23.684 "uuid": "b1d9ac8a-6359-4630-98b7-374f84ba7728", 00:10:23.684 "is_configured": true, 00:10:23.684 "data_offset": 2048, 00:10:23.684 "data_size": 63488 00:10:23.684 } 00:10:23.684 ] 00:10:23.684 }' 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:23.684 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:23.944 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:23.944 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:23.944 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:23.944 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:23.944 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:23.944 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:23.944 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:23.944 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:23.944 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:23.944 [2024-11-27 15:14:51.988034] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:23.944 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:23.944 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:23.944 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:23.944 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:23.944 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:23.944 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:23.944 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:23.944 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:23.944 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:23.944 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:23.944 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:23.944 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:23.944 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:23.944 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:23.944 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:23.944 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:23.944 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:23.944 "name": "Existed_Raid", 00:10:23.944 "uuid": "27c11bfe-147c-47e8-82a2-0db8e51ba085", 00:10:23.944 "strip_size_kb": 64, 00:10:23.944 "state": "configuring", 00:10:23.944 "raid_level": "concat", 00:10:23.944 "superblock": true, 00:10:23.944 "num_base_bdevs": 4, 00:10:23.944 "num_base_bdevs_discovered": 2, 00:10:23.944 "num_base_bdevs_operational": 4, 00:10:23.944 "base_bdevs_list": [ 00:10:23.944 { 00:10:23.944 "name": "BaseBdev1", 00:10:23.944 "uuid": "2b95f3ea-5e40-4960-91d3-0105f9af776c", 00:10:23.944 "is_configured": true, 00:10:23.944 "data_offset": 2048, 00:10:23.944 "data_size": 63488 00:10:23.944 }, 00:10:23.944 { 00:10:23.944 "name": null, 00:10:23.944 "uuid": "ddb0aa0b-7ed5-4ae2-b6a7-cb713fa28ae8", 00:10:23.944 "is_configured": false, 00:10:23.944 "data_offset": 0, 00:10:23.944 "data_size": 63488 00:10:23.944 }, 00:10:23.944 { 00:10:23.944 "name": null, 00:10:23.944 "uuid": "82d2514e-25e5-4f74-8790-f575e2ca382a", 00:10:23.944 "is_configured": false, 00:10:23.944 "data_offset": 0, 00:10:23.944 "data_size": 63488 00:10:23.944 }, 00:10:23.944 { 00:10:23.944 "name": "BaseBdev4", 00:10:23.944 "uuid": "b1d9ac8a-6359-4630-98b7-374f84ba7728", 00:10:23.944 "is_configured": true, 00:10:23.944 "data_offset": 2048, 00:10:23.944 "data_size": 63488 00:10:23.944 } 00:10:23.944 ] 00:10:23.944 }' 00:10:23.944 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:23.944 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:24.514 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:24.514 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:24.514 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:24.514 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:24.514 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:24.514 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:24.514 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:24.514 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:24.514 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:24.514 [2024-11-27 15:14:52.451403] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:24.514 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:24.514 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:24.514 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:24.514 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:24.514 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:24.514 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:24.514 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:24.514 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:24.514 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:24.514 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:24.514 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:24.514 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:24.514 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:24.514 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:24.514 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:24.514 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:24.514 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:24.514 "name": "Existed_Raid", 00:10:24.514 "uuid": "27c11bfe-147c-47e8-82a2-0db8e51ba085", 00:10:24.514 "strip_size_kb": 64, 00:10:24.514 "state": "configuring", 00:10:24.514 "raid_level": "concat", 00:10:24.514 "superblock": true, 00:10:24.514 "num_base_bdevs": 4, 00:10:24.514 "num_base_bdevs_discovered": 3, 00:10:24.514 "num_base_bdevs_operational": 4, 00:10:24.514 "base_bdevs_list": [ 00:10:24.514 { 00:10:24.514 "name": "BaseBdev1", 00:10:24.514 "uuid": "2b95f3ea-5e40-4960-91d3-0105f9af776c", 00:10:24.514 "is_configured": true, 00:10:24.514 "data_offset": 2048, 00:10:24.514 "data_size": 63488 00:10:24.514 }, 00:10:24.514 { 00:10:24.514 "name": null, 00:10:24.514 "uuid": "ddb0aa0b-7ed5-4ae2-b6a7-cb713fa28ae8", 00:10:24.514 "is_configured": false, 00:10:24.514 "data_offset": 0, 00:10:24.514 "data_size": 63488 00:10:24.514 }, 00:10:24.514 { 00:10:24.514 "name": "BaseBdev3", 00:10:24.514 "uuid": "82d2514e-25e5-4f74-8790-f575e2ca382a", 00:10:24.514 "is_configured": true, 00:10:24.514 "data_offset": 2048, 00:10:24.514 "data_size": 63488 00:10:24.514 }, 00:10:24.514 { 00:10:24.514 "name": "BaseBdev4", 00:10:24.514 "uuid": "b1d9ac8a-6359-4630-98b7-374f84ba7728", 00:10:24.514 "is_configured": true, 00:10:24.514 "data_offset": 2048, 00:10:24.514 "data_size": 63488 00:10:24.514 } 00:10:24.514 ] 00:10:24.514 }' 00:10:24.514 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:24.514 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.084 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:25.084 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.084 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.084 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:25.084 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.084 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:25.084 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:25.084 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.085 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.085 [2024-11-27 15:14:52.958487] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:25.085 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.085 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:25.085 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:25.085 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:25.085 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:25.085 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:25.085 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:25.085 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:25.085 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:25.085 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:25.085 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:25.085 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:25.085 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.085 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.085 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:25.085 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.085 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:25.085 "name": "Existed_Raid", 00:10:25.085 "uuid": "27c11bfe-147c-47e8-82a2-0db8e51ba085", 00:10:25.085 "strip_size_kb": 64, 00:10:25.085 "state": "configuring", 00:10:25.085 "raid_level": "concat", 00:10:25.085 "superblock": true, 00:10:25.085 "num_base_bdevs": 4, 00:10:25.085 "num_base_bdevs_discovered": 2, 00:10:25.085 "num_base_bdevs_operational": 4, 00:10:25.085 "base_bdevs_list": [ 00:10:25.085 { 00:10:25.085 "name": null, 00:10:25.085 "uuid": "2b95f3ea-5e40-4960-91d3-0105f9af776c", 00:10:25.085 "is_configured": false, 00:10:25.085 "data_offset": 0, 00:10:25.085 "data_size": 63488 00:10:25.085 }, 00:10:25.085 { 00:10:25.085 "name": null, 00:10:25.085 "uuid": "ddb0aa0b-7ed5-4ae2-b6a7-cb713fa28ae8", 00:10:25.085 "is_configured": false, 00:10:25.085 "data_offset": 0, 00:10:25.085 "data_size": 63488 00:10:25.085 }, 00:10:25.085 { 00:10:25.085 "name": "BaseBdev3", 00:10:25.085 "uuid": "82d2514e-25e5-4f74-8790-f575e2ca382a", 00:10:25.085 "is_configured": true, 00:10:25.085 "data_offset": 2048, 00:10:25.085 "data_size": 63488 00:10:25.085 }, 00:10:25.085 { 00:10:25.085 "name": "BaseBdev4", 00:10:25.085 "uuid": "b1d9ac8a-6359-4630-98b7-374f84ba7728", 00:10:25.085 "is_configured": true, 00:10:25.085 "data_offset": 2048, 00:10:25.085 "data_size": 63488 00:10:25.085 } 00:10:25.085 ] 00:10:25.085 }' 00:10:25.085 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:25.085 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.344 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:25.344 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:25.344 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.344 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.344 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.344 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:25.344 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:25.344 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.344 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.344 [2024-11-27 15:14:53.428151] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:25.344 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.344 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:25.344 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:25.344 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:25.344 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:25.344 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:25.344 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:25.344 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:25.344 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:25.344 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:25.344 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:25.344 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:25.344 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:25.344 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.344 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.604 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.604 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:25.604 "name": "Existed_Raid", 00:10:25.604 "uuid": "27c11bfe-147c-47e8-82a2-0db8e51ba085", 00:10:25.604 "strip_size_kb": 64, 00:10:25.604 "state": "configuring", 00:10:25.604 "raid_level": "concat", 00:10:25.604 "superblock": true, 00:10:25.604 "num_base_bdevs": 4, 00:10:25.604 "num_base_bdevs_discovered": 3, 00:10:25.604 "num_base_bdevs_operational": 4, 00:10:25.604 "base_bdevs_list": [ 00:10:25.604 { 00:10:25.604 "name": null, 00:10:25.604 "uuid": "2b95f3ea-5e40-4960-91d3-0105f9af776c", 00:10:25.604 "is_configured": false, 00:10:25.604 "data_offset": 0, 00:10:25.604 "data_size": 63488 00:10:25.604 }, 00:10:25.604 { 00:10:25.604 "name": "BaseBdev2", 00:10:25.604 "uuid": "ddb0aa0b-7ed5-4ae2-b6a7-cb713fa28ae8", 00:10:25.604 "is_configured": true, 00:10:25.604 "data_offset": 2048, 00:10:25.604 "data_size": 63488 00:10:25.604 }, 00:10:25.604 { 00:10:25.604 "name": "BaseBdev3", 00:10:25.604 "uuid": "82d2514e-25e5-4f74-8790-f575e2ca382a", 00:10:25.604 "is_configured": true, 00:10:25.604 "data_offset": 2048, 00:10:25.604 "data_size": 63488 00:10:25.604 }, 00:10:25.604 { 00:10:25.604 "name": "BaseBdev4", 00:10:25.604 "uuid": "b1d9ac8a-6359-4630-98b7-374f84ba7728", 00:10:25.604 "is_configured": true, 00:10:25.604 "data_offset": 2048, 00:10:25.604 "data_size": 63488 00:10:25.604 } 00:10:25.604 ] 00:10:25.604 }' 00:10:25.604 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:25.604 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.864 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:25.864 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:25.864 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.864 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.864 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.864 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:25.864 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:25.864 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:25.864 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.864 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.864 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.864 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 2b95f3ea-5e40-4960-91d3-0105f9af776c 00:10:25.864 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.864 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.864 [2024-11-27 15:14:53.954294] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:25.864 [2024-11-27 15:14:53.954540] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:10:25.864 [2024-11-27 15:14:53.954577] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:25.864 NewBaseBdev 00:10:25.864 [2024-11-27 15:14:53.954873] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:10:25.864 [2024-11-27 15:14:53.955009] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:10:25.864 [2024-11-27 15:14:53.955062] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006d00 00:10:25.864 [2024-11-27 15:14:53.955198] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:25.864 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.864 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:25.864 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:10:25.864 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:25.864 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:25.864 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:25.864 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:25.864 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:25.864 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.864 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.864 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.864 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:25.864 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.864 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.123 [ 00:10:26.123 { 00:10:26.123 "name": "NewBaseBdev", 00:10:26.123 "aliases": [ 00:10:26.123 "2b95f3ea-5e40-4960-91d3-0105f9af776c" 00:10:26.123 ], 00:10:26.123 "product_name": "Malloc disk", 00:10:26.123 "block_size": 512, 00:10:26.123 "num_blocks": 65536, 00:10:26.123 "uuid": "2b95f3ea-5e40-4960-91d3-0105f9af776c", 00:10:26.123 "assigned_rate_limits": { 00:10:26.123 "rw_ios_per_sec": 0, 00:10:26.123 "rw_mbytes_per_sec": 0, 00:10:26.123 "r_mbytes_per_sec": 0, 00:10:26.123 "w_mbytes_per_sec": 0 00:10:26.123 }, 00:10:26.123 "claimed": true, 00:10:26.123 "claim_type": "exclusive_write", 00:10:26.123 "zoned": false, 00:10:26.123 "supported_io_types": { 00:10:26.123 "read": true, 00:10:26.123 "write": true, 00:10:26.123 "unmap": true, 00:10:26.123 "flush": true, 00:10:26.123 "reset": true, 00:10:26.123 "nvme_admin": false, 00:10:26.123 "nvme_io": false, 00:10:26.123 "nvme_io_md": false, 00:10:26.123 "write_zeroes": true, 00:10:26.123 "zcopy": true, 00:10:26.123 "get_zone_info": false, 00:10:26.123 "zone_management": false, 00:10:26.123 "zone_append": false, 00:10:26.123 "compare": false, 00:10:26.123 "compare_and_write": false, 00:10:26.123 "abort": true, 00:10:26.123 "seek_hole": false, 00:10:26.123 "seek_data": false, 00:10:26.123 "copy": true, 00:10:26.123 "nvme_iov_md": false 00:10:26.123 }, 00:10:26.123 "memory_domains": [ 00:10:26.123 { 00:10:26.123 "dma_device_id": "system", 00:10:26.123 "dma_device_type": 1 00:10:26.123 }, 00:10:26.123 { 00:10:26.123 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:26.124 "dma_device_type": 2 00:10:26.124 } 00:10:26.124 ], 00:10:26.124 "driver_specific": {} 00:10:26.124 } 00:10:26.124 ] 00:10:26.124 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.124 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:26.124 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:10:26.124 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:26.124 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:26.124 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:26.124 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:26.124 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:26.124 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:26.124 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:26.124 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:26.124 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:26.124 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:26.124 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:26.124 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.124 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.124 15:14:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.124 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:26.124 "name": "Existed_Raid", 00:10:26.124 "uuid": "27c11bfe-147c-47e8-82a2-0db8e51ba085", 00:10:26.124 "strip_size_kb": 64, 00:10:26.124 "state": "online", 00:10:26.124 "raid_level": "concat", 00:10:26.124 "superblock": true, 00:10:26.124 "num_base_bdevs": 4, 00:10:26.124 "num_base_bdevs_discovered": 4, 00:10:26.124 "num_base_bdevs_operational": 4, 00:10:26.124 "base_bdevs_list": [ 00:10:26.124 { 00:10:26.124 "name": "NewBaseBdev", 00:10:26.124 "uuid": "2b95f3ea-5e40-4960-91d3-0105f9af776c", 00:10:26.124 "is_configured": true, 00:10:26.124 "data_offset": 2048, 00:10:26.124 "data_size": 63488 00:10:26.124 }, 00:10:26.124 { 00:10:26.124 "name": "BaseBdev2", 00:10:26.124 "uuid": "ddb0aa0b-7ed5-4ae2-b6a7-cb713fa28ae8", 00:10:26.124 "is_configured": true, 00:10:26.124 "data_offset": 2048, 00:10:26.124 "data_size": 63488 00:10:26.124 }, 00:10:26.124 { 00:10:26.124 "name": "BaseBdev3", 00:10:26.124 "uuid": "82d2514e-25e5-4f74-8790-f575e2ca382a", 00:10:26.124 "is_configured": true, 00:10:26.124 "data_offset": 2048, 00:10:26.124 "data_size": 63488 00:10:26.124 }, 00:10:26.124 { 00:10:26.124 "name": "BaseBdev4", 00:10:26.124 "uuid": "b1d9ac8a-6359-4630-98b7-374f84ba7728", 00:10:26.124 "is_configured": true, 00:10:26.124 "data_offset": 2048, 00:10:26.124 "data_size": 63488 00:10:26.124 } 00:10:26.124 ] 00:10:26.124 }' 00:10:26.124 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:26.124 15:14:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.384 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:26.384 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:26.384 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:26.384 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:26.384 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:26.384 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:26.384 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:26.384 15:14:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.384 15:14:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.384 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:26.384 [2024-11-27 15:14:54.322035] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:26.384 15:14:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.384 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:26.384 "name": "Existed_Raid", 00:10:26.384 "aliases": [ 00:10:26.384 "27c11bfe-147c-47e8-82a2-0db8e51ba085" 00:10:26.384 ], 00:10:26.384 "product_name": "Raid Volume", 00:10:26.384 "block_size": 512, 00:10:26.384 "num_blocks": 253952, 00:10:26.384 "uuid": "27c11bfe-147c-47e8-82a2-0db8e51ba085", 00:10:26.384 "assigned_rate_limits": { 00:10:26.384 "rw_ios_per_sec": 0, 00:10:26.384 "rw_mbytes_per_sec": 0, 00:10:26.384 "r_mbytes_per_sec": 0, 00:10:26.384 "w_mbytes_per_sec": 0 00:10:26.384 }, 00:10:26.384 "claimed": false, 00:10:26.384 "zoned": false, 00:10:26.384 "supported_io_types": { 00:10:26.384 "read": true, 00:10:26.384 "write": true, 00:10:26.384 "unmap": true, 00:10:26.384 "flush": true, 00:10:26.384 "reset": true, 00:10:26.384 "nvme_admin": false, 00:10:26.384 "nvme_io": false, 00:10:26.384 "nvme_io_md": false, 00:10:26.384 "write_zeroes": true, 00:10:26.384 "zcopy": false, 00:10:26.384 "get_zone_info": false, 00:10:26.384 "zone_management": false, 00:10:26.384 "zone_append": false, 00:10:26.384 "compare": false, 00:10:26.384 "compare_and_write": false, 00:10:26.384 "abort": false, 00:10:26.384 "seek_hole": false, 00:10:26.384 "seek_data": false, 00:10:26.384 "copy": false, 00:10:26.384 "nvme_iov_md": false 00:10:26.384 }, 00:10:26.384 "memory_domains": [ 00:10:26.384 { 00:10:26.384 "dma_device_id": "system", 00:10:26.384 "dma_device_type": 1 00:10:26.384 }, 00:10:26.384 { 00:10:26.384 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:26.384 "dma_device_type": 2 00:10:26.384 }, 00:10:26.384 { 00:10:26.384 "dma_device_id": "system", 00:10:26.384 "dma_device_type": 1 00:10:26.384 }, 00:10:26.384 { 00:10:26.384 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:26.384 "dma_device_type": 2 00:10:26.384 }, 00:10:26.384 { 00:10:26.384 "dma_device_id": "system", 00:10:26.384 "dma_device_type": 1 00:10:26.384 }, 00:10:26.384 { 00:10:26.384 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:26.384 "dma_device_type": 2 00:10:26.384 }, 00:10:26.384 { 00:10:26.384 "dma_device_id": "system", 00:10:26.384 "dma_device_type": 1 00:10:26.384 }, 00:10:26.384 { 00:10:26.384 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:26.384 "dma_device_type": 2 00:10:26.384 } 00:10:26.384 ], 00:10:26.384 "driver_specific": { 00:10:26.384 "raid": { 00:10:26.384 "uuid": "27c11bfe-147c-47e8-82a2-0db8e51ba085", 00:10:26.384 "strip_size_kb": 64, 00:10:26.384 "state": "online", 00:10:26.384 "raid_level": "concat", 00:10:26.384 "superblock": true, 00:10:26.384 "num_base_bdevs": 4, 00:10:26.384 "num_base_bdevs_discovered": 4, 00:10:26.384 "num_base_bdevs_operational": 4, 00:10:26.384 "base_bdevs_list": [ 00:10:26.384 { 00:10:26.384 "name": "NewBaseBdev", 00:10:26.384 "uuid": "2b95f3ea-5e40-4960-91d3-0105f9af776c", 00:10:26.384 "is_configured": true, 00:10:26.384 "data_offset": 2048, 00:10:26.384 "data_size": 63488 00:10:26.384 }, 00:10:26.384 { 00:10:26.384 "name": "BaseBdev2", 00:10:26.384 "uuid": "ddb0aa0b-7ed5-4ae2-b6a7-cb713fa28ae8", 00:10:26.384 "is_configured": true, 00:10:26.384 "data_offset": 2048, 00:10:26.384 "data_size": 63488 00:10:26.384 }, 00:10:26.384 { 00:10:26.384 "name": "BaseBdev3", 00:10:26.384 "uuid": "82d2514e-25e5-4f74-8790-f575e2ca382a", 00:10:26.384 "is_configured": true, 00:10:26.384 "data_offset": 2048, 00:10:26.384 "data_size": 63488 00:10:26.384 }, 00:10:26.384 { 00:10:26.384 "name": "BaseBdev4", 00:10:26.384 "uuid": "b1d9ac8a-6359-4630-98b7-374f84ba7728", 00:10:26.384 "is_configured": true, 00:10:26.384 "data_offset": 2048, 00:10:26.384 "data_size": 63488 00:10:26.384 } 00:10:26.384 ] 00:10:26.384 } 00:10:26.384 } 00:10:26.384 }' 00:10:26.384 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:26.384 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:26.384 BaseBdev2 00:10:26.384 BaseBdev3 00:10:26.384 BaseBdev4' 00:10:26.384 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:26.384 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:26.384 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:26.384 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:26.384 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:26.384 15:14:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.384 15:14:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.384 15:14:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.645 [2024-11-27 15:14:54.657157] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:26.645 [2024-11-27 15:14:54.657199] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:26.645 [2024-11-27 15:14:54.657288] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:26.645 [2024-11-27 15:14:54.657359] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:26.645 [2024-11-27 15:14:54.657369] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name Existed_Raid, state offline 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 82970 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 82970 ']' 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 82970 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 82970 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:26.645 killing process with pid 82970 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 82970' 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 82970 00:10:26.645 [2024-11-27 15:14:54.698269] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:26.645 15:14:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 82970 00:10:26.645 [2024-11-27 15:14:54.739855] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:26.906 15:14:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:10:26.906 00:10:26.906 real 0m9.172s 00:10:26.906 user 0m15.548s 00:10:26.906 sys 0m2.097s 00:10:26.906 ************************************ 00:10:26.906 END TEST raid_state_function_test_sb 00:10:26.906 ************************************ 00:10:26.906 15:14:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:26.906 15:14:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:27.166 15:14:55 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 4 00:10:27.166 15:14:55 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:10:27.166 15:14:55 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:27.166 15:14:55 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:27.166 ************************************ 00:10:27.166 START TEST raid_superblock_test 00:10:27.166 ************************************ 00:10:27.166 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test concat 4 00:10:27.166 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:10:27.166 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:10:27.166 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:10:27.166 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:10:27.166 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:10:27.166 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:10:27.166 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:10:27.166 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:10:27.166 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:10:27.166 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:10:27.166 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:10:27.166 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:10:27.166 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:10:27.166 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:10:27.166 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:10:27.166 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:10:27.166 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=83617 00:10:27.166 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 83617 00:10:27.166 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 83617 ']' 00:10:27.166 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:10:27.166 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:27.166 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:27.166 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:27.166 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:27.166 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:27.166 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.166 [2024-11-27 15:14:55.139959] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:10:27.166 [2024-11-27 15:14:55.140209] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid83617 ] 00:10:27.426 [2024-11-27 15:14:55.314236] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:27.426 [2024-11-27 15:14:55.340029] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:27.426 [2024-11-27 15:14:55.382953] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:27.426 [2024-11-27 15:14:55.382992] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:27.995 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:27.995 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:10:27.995 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:10:27.995 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:27.995 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:10:27.995 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:10:27.996 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:10:27.996 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:27.996 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:27.996 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:27.996 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:10:27.996 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.996 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.996 malloc1 00:10:27.996 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.996 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:27.996 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.996 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.996 [2024-11-27 15:14:55.990893] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:27.996 [2024-11-27 15:14:55.991032] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:27.996 [2024-11-27 15:14:55.991072] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:10:27.996 [2024-11-27 15:14:55.991111] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:27.996 [2024-11-27 15:14:55.993340] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:27.996 [2024-11-27 15:14:55.993435] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:27.996 pt1 00:10:27.996 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.996 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:27.996 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:27.996 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:10:27.996 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:10:27.996 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:10:27.996 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:27.996 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:27.996 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:27.996 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:10:27.996 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.996 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.996 malloc2 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.996 [2024-11-27 15:14:56.019528] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:27.996 [2024-11-27 15:14:56.019660] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:27.996 [2024-11-27 15:14:56.019695] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:10:27.996 [2024-11-27 15:14:56.019731] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:27.996 [2024-11-27 15:14:56.021845] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:27.996 [2024-11-27 15:14:56.021932] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:27.996 pt2 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.996 malloc3 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.996 [2024-11-27 15:14:56.048289] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:27.996 [2024-11-27 15:14:56.048415] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:27.996 [2024-11-27 15:14:56.048456] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:10:27.996 [2024-11-27 15:14:56.048509] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:27.996 [2024-11-27 15:14:56.050682] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:27.996 [2024-11-27 15:14:56.050759] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:27.996 pt3 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.996 malloc4 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.996 [2024-11-27 15:14:56.092748] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:27.996 [2024-11-27 15:14:56.092853] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:27.996 [2024-11-27 15:14:56.092890] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:27.996 [2024-11-27 15:14:56.092941] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:27.996 [2024-11-27 15:14:56.095061] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:27.996 [2024-11-27 15:14:56.095150] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:27.996 pt4 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.996 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:28.256 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:28.256 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:10:28.256 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.256 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.256 [2024-11-27 15:14:56.104798] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:28.256 [2024-11-27 15:14:56.106685] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:28.256 [2024-11-27 15:14:56.106764] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:28.256 [2024-11-27 15:14:56.106811] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:28.256 [2024-11-27 15:14:56.106970] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:10:28.256 [2024-11-27 15:14:56.106988] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:28.256 [2024-11-27 15:14:56.107253] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:10:28.256 [2024-11-27 15:14:56.107400] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:10:28.256 [2024-11-27 15:14:56.107410] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:10:28.256 [2024-11-27 15:14:56.107534] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:28.256 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.256 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:28.256 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:28.256 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:28.256 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:28.256 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:28.256 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:28.256 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:28.256 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:28.256 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:28.256 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:28.256 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:28.256 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:28.256 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.256 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.256 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.256 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:28.256 "name": "raid_bdev1", 00:10:28.256 "uuid": "7efd1cef-d739-40ca-a820-d086958ecd0d", 00:10:28.256 "strip_size_kb": 64, 00:10:28.256 "state": "online", 00:10:28.256 "raid_level": "concat", 00:10:28.256 "superblock": true, 00:10:28.256 "num_base_bdevs": 4, 00:10:28.256 "num_base_bdevs_discovered": 4, 00:10:28.256 "num_base_bdevs_operational": 4, 00:10:28.256 "base_bdevs_list": [ 00:10:28.256 { 00:10:28.256 "name": "pt1", 00:10:28.256 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:28.256 "is_configured": true, 00:10:28.256 "data_offset": 2048, 00:10:28.256 "data_size": 63488 00:10:28.256 }, 00:10:28.256 { 00:10:28.256 "name": "pt2", 00:10:28.256 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:28.256 "is_configured": true, 00:10:28.256 "data_offset": 2048, 00:10:28.256 "data_size": 63488 00:10:28.256 }, 00:10:28.256 { 00:10:28.256 "name": "pt3", 00:10:28.256 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:28.256 "is_configured": true, 00:10:28.256 "data_offset": 2048, 00:10:28.256 "data_size": 63488 00:10:28.256 }, 00:10:28.256 { 00:10:28.256 "name": "pt4", 00:10:28.256 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:28.256 "is_configured": true, 00:10:28.257 "data_offset": 2048, 00:10:28.257 "data_size": 63488 00:10:28.257 } 00:10:28.257 ] 00:10:28.257 }' 00:10:28.257 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:28.257 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.516 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:10:28.516 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:28.516 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:28.516 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:28.516 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:28.516 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:28.516 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:28.516 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:28.516 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.516 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.516 [2024-11-27 15:14:56.568326] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:28.516 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.516 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:28.517 "name": "raid_bdev1", 00:10:28.517 "aliases": [ 00:10:28.517 "7efd1cef-d739-40ca-a820-d086958ecd0d" 00:10:28.517 ], 00:10:28.517 "product_name": "Raid Volume", 00:10:28.517 "block_size": 512, 00:10:28.517 "num_blocks": 253952, 00:10:28.517 "uuid": "7efd1cef-d739-40ca-a820-d086958ecd0d", 00:10:28.517 "assigned_rate_limits": { 00:10:28.517 "rw_ios_per_sec": 0, 00:10:28.517 "rw_mbytes_per_sec": 0, 00:10:28.517 "r_mbytes_per_sec": 0, 00:10:28.517 "w_mbytes_per_sec": 0 00:10:28.517 }, 00:10:28.517 "claimed": false, 00:10:28.517 "zoned": false, 00:10:28.517 "supported_io_types": { 00:10:28.517 "read": true, 00:10:28.517 "write": true, 00:10:28.517 "unmap": true, 00:10:28.517 "flush": true, 00:10:28.517 "reset": true, 00:10:28.517 "nvme_admin": false, 00:10:28.517 "nvme_io": false, 00:10:28.517 "nvme_io_md": false, 00:10:28.517 "write_zeroes": true, 00:10:28.517 "zcopy": false, 00:10:28.517 "get_zone_info": false, 00:10:28.517 "zone_management": false, 00:10:28.517 "zone_append": false, 00:10:28.517 "compare": false, 00:10:28.517 "compare_and_write": false, 00:10:28.517 "abort": false, 00:10:28.517 "seek_hole": false, 00:10:28.517 "seek_data": false, 00:10:28.517 "copy": false, 00:10:28.517 "nvme_iov_md": false 00:10:28.517 }, 00:10:28.517 "memory_domains": [ 00:10:28.517 { 00:10:28.517 "dma_device_id": "system", 00:10:28.517 "dma_device_type": 1 00:10:28.517 }, 00:10:28.517 { 00:10:28.517 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:28.517 "dma_device_type": 2 00:10:28.517 }, 00:10:28.517 { 00:10:28.517 "dma_device_id": "system", 00:10:28.517 "dma_device_type": 1 00:10:28.517 }, 00:10:28.517 { 00:10:28.517 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:28.517 "dma_device_type": 2 00:10:28.517 }, 00:10:28.517 { 00:10:28.517 "dma_device_id": "system", 00:10:28.517 "dma_device_type": 1 00:10:28.517 }, 00:10:28.517 { 00:10:28.517 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:28.517 "dma_device_type": 2 00:10:28.517 }, 00:10:28.517 { 00:10:28.517 "dma_device_id": "system", 00:10:28.517 "dma_device_type": 1 00:10:28.517 }, 00:10:28.517 { 00:10:28.517 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:28.517 "dma_device_type": 2 00:10:28.517 } 00:10:28.517 ], 00:10:28.517 "driver_specific": { 00:10:28.517 "raid": { 00:10:28.517 "uuid": "7efd1cef-d739-40ca-a820-d086958ecd0d", 00:10:28.517 "strip_size_kb": 64, 00:10:28.517 "state": "online", 00:10:28.517 "raid_level": "concat", 00:10:28.517 "superblock": true, 00:10:28.517 "num_base_bdevs": 4, 00:10:28.517 "num_base_bdevs_discovered": 4, 00:10:28.517 "num_base_bdevs_operational": 4, 00:10:28.517 "base_bdevs_list": [ 00:10:28.517 { 00:10:28.517 "name": "pt1", 00:10:28.517 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:28.517 "is_configured": true, 00:10:28.517 "data_offset": 2048, 00:10:28.517 "data_size": 63488 00:10:28.517 }, 00:10:28.517 { 00:10:28.517 "name": "pt2", 00:10:28.517 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:28.517 "is_configured": true, 00:10:28.517 "data_offset": 2048, 00:10:28.517 "data_size": 63488 00:10:28.517 }, 00:10:28.517 { 00:10:28.517 "name": "pt3", 00:10:28.517 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:28.517 "is_configured": true, 00:10:28.517 "data_offset": 2048, 00:10:28.517 "data_size": 63488 00:10:28.517 }, 00:10:28.517 { 00:10:28.517 "name": "pt4", 00:10:28.517 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:28.517 "is_configured": true, 00:10:28.517 "data_offset": 2048, 00:10:28.517 "data_size": 63488 00:10:28.517 } 00:10:28.517 ] 00:10:28.517 } 00:10:28.517 } 00:10:28.517 }' 00:10:28.517 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:28.776 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:28.776 pt2 00:10:28.776 pt3 00:10:28.776 pt4' 00:10:28.776 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:28.776 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:28.776 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:28.776 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:28.776 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:28.776 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.776 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.776 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.776 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:28.776 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:28.776 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:28.776 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:28.776 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.776 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.776 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:28.776 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.776 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:28.776 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:28.776 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:28.776 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:28.777 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.777 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.777 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:28.777 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.777 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:28.777 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:28.777 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:28.777 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:28.777 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:10:28.777 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.777 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.777 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.777 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:28.777 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:28.777 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:10:28.777 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:28.777 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.777 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.777 [2024-11-27 15:14:56.871842] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:29.037 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.037 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=7efd1cef-d739-40ca-a820-d086958ecd0d 00:10:29.037 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 7efd1cef-d739-40ca-a820-d086958ecd0d ']' 00:10:29.037 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:29.037 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.037 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.037 [2024-11-27 15:14:56.907557] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:29.037 [2024-11-27 15:14:56.907670] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:29.037 [2024-11-27 15:14:56.907781] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:29.037 [2024-11-27 15:14:56.907884] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:29.037 [2024-11-27 15:14:56.907953] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:10:29.037 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.037 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:29.037 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.037 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:10:29.037 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.037 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.037 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:10:29.037 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:10:29.037 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:29.037 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:10:29.037 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.037 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.037 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.037 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:29.037 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:10:29.037 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.037 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.037 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.037 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:29.037 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:10:29.038 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.038 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.038 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.038 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:29.038 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:10:29.038 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.038 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.038 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.038 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:10:29.038 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.038 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.038 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.038 [2024-11-27 15:14:57.051330] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:10:29.038 [2024-11-27 15:14:57.053333] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:10:29.038 [2024-11-27 15:14:57.053431] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:10:29.038 [2024-11-27 15:14:57.053479] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:10:29.038 [2024-11-27 15:14:57.053543] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:10:29.038 [2024-11-27 15:14:57.053650] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:10:29.038 [2024-11-27 15:14:57.053708] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:10:29.038 [2024-11-27 15:14:57.053824] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:10:29.038 [2024-11-27 15:14:57.053890] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:29.038 [2024-11-27 15:14:57.053935] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state configuring 00:10:29.038 request: 00:10:29.038 { 00:10:29.038 "name": "raid_bdev1", 00:10:29.038 "raid_level": "concat", 00:10:29.038 "base_bdevs": [ 00:10:29.038 "malloc1", 00:10:29.038 "malloc2", 00:10:29.038 "malloc3", 00:10:29.038 "malloc4" 00:10:29.038 ], 00:10:29.038 "strip_size_kb": 64, 00:10:29.038 "superblock": false, 00:10:29.038 "method": "bdev_raid_create", 00:10:29.038 "req_id": 1 00:10:29.038 } 00:10:29.038 Got JSON-RPC error response 00:10:29.038 response: 00:10:29.038 { 00:10:29.038 "code": -17, 00:10:29.038 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:10:29.038 } 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.038 [2024-11-27 15:14:57.107170] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:29.038 [2024-11-27 15:14:57.107311] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:29.038 [2024-11-27 15:14:57.107350] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:10:29.038 [2024-11-27 15:14:57.107380] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:29.038 [2024-11-27 15:14:57.109559] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:29.038 [2024-11-27 15:14:57.109639] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:29.038 [2024-11-27 15:14:57.109760] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:10:29.038 [2024-11-27 15:14:57.109836] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:29.038 pt1 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 4 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.038 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.298 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:29.298 "name": "raid_bdev1", 00:10:29.298 "uuid": "7efd1cef-d739-40ca-a820-d086958ecd0d", 00:10:29.298 "strip_size_kb": 64, 00:10:29.298 "state": "configuring", 00:10:29.298 "raid_level": "concat", 00:10:29.298 "superblock": true, 00:10:29.298 "num_base_bdevs": 4, 00:10:29.298 "num_base_bdevs_discovered": 1, 00:10:29.298 "num_base_bdevs_operational": 4, 00:10:29.298 "base_bdevs_list": [ 00:10:29.298 { 00:10:29.298 "name": "pt1", 00:10:29.298 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:29.298 "is_configured": true, 00:10:29.298 "data_offset": 2048, 00:10:29.298 "data_size": 63488 00:10:29.298 }, 00:10:29.298 { 00:10:29.298 "name": null, 00:10:29.298 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:29.298 "is_configured": false, 00:10:29.298 "data_offset": 2048, 00:10:29.299 "data_size": 63488 00:10:29.299 }, 00:10:29.299 { 00:10:29.299 "name": null, 00:10:29.299 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:29.299 "is_configured": false, 00:10:29.299 "data_offset": 2048, 00:10:29.299 "data_size": 63488 00:10:29.299 }, 00:10:29.299 { 00:10:29.299 "name": null, 00:10:29.299 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:29.299 "is_configured": false, 00:10:29.299 "data_offset": 2048, 00:10:29.299 "data_size": 63488 00:10:29.299 } 00:10:29.299 ] 00:10:29.299 }' 00:10:29.299 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:29.299 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.559 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:10:29.559 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:29.559 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.559 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.559 [2024-11-27 15:14:57.574360] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:29.559 [2024-11-27 15:14:57.574460] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:29.559 [2024-11-27 15:14:57.574508] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:10:29.559 [2024-11-27 15:14:57.574539] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:29.559 [2024-11-27 15:14:57.574958] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:29.559 [2024-11-27 15:14:57.575021] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:29.559 [2024-11-27 15:14:57.575130] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:29.559 [2024-11-27 15:14:57.575180] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:29.559 pt2 00:10:29.559 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.559 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:10:29.559 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.559 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.559 [2024-11-27 15:14:57.586339] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:10:29.559 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.559 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 4 00:10:29.559 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:29.559 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:29.559 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:29.559 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:29.559 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:29.559 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:29.559 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:29.559 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:29.559 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:29.559 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:29.559 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:29.559 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.559 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.559 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.559 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:29.559 "name": "raid_bdev1", 00:10:29.559 "uuid": "7efd1cef-d739-40ca-a820-d086958ecd0d", 00:10:29.559 "strip_size_kb": 64, 00:10:29.559 "state": "configuring", 00:10:29.559 "raid_level": "concat", 00:10:29.559 "superblock": true, 00:10:29.559 "num_base_bdevs": 4, 00:10:29.559 "num_base_bdevs_discovered": 1, 00:10:29.559 "num_base_bdevs_operational": 4, 00:10:29.559 "base_bdevs_list": [ 00:10:29.559 { 00:10:29.559 "name": "pt1", 00:10:29.559 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:29.559 "is_configured": true, 00:10:29.559 "data_offset": 2048, 00:10:29.559 "data_size": 63488 00:10:29.559 }, 00:10:29.559 { 00:10:29.559 "name": null, 00:10:29.559 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:29.559 "is_configured": false, 00:10:29.559 "data_offset": 0, 00:10:29.559 "data_size": 63488 00:10:29.559 }, 00:10:29.559 { 00:10:29.559 "name": null, 00:10:29.559 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:29.559 "is_configured": false, 00:10:29.559 "data_offset": 2048, 00:10:29.559 "data_size": 63488 00:10:29.559 }, 00:10:29.559 { 00:10:29.559 "name": null, 00:10:29.559 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:29.559 "is_configured": false, 00:10:29.559 "data_offset": 2048, 00:10:29.559 "data_size": 63488 00:10:29.559 } 00:10:29.559 ] 00:10:29.559 }' 00:10:29.559 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:29.559 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.139 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:10:30.139 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:30.139 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:30.139 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.139 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.139 [2024-11-27 15:14:58.029577] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:30.139 [2024-11-27 15:14:58.029705] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:30.139 [2024-11-27 15:14:58.029726] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:10:30.139 [2024-11-27 15:14:58.029737] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:30.139 [2024-11-27 15:14:58.030133] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:30.139 [2024-11-27 15:14:58.030162] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:30.139 [2024-11-27 15:14:58.030235] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:30.139 [2024-11-27 15:14:58.030259] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:30.139 pt2 00:10:30.139 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.139 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:30.139 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:30.139 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:30.139 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.139 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.139 [2024-11-27 15:14:58.041548] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:30.140 [2024-11-27 15:14:58.041599] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:30.140 [2024-11-27 15:14:58.041619] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:10:30.140 [2024-11-27 15:14:58.041632] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:30.140 [2024-11-27 15:14:58.041956] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:30.140 [2024-11-27 15:14:58.041978] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:30.140 [2024-11-27 15:14:58.042031] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:10:30.140 [2024-11-27 15:14:58.042062] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:30.140 pt3 00:10:30.140 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.140 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:30.140 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:30.140 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:30.140 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.140 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.140 [2024-11-27 15:14:58.053496] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:30.140 [2024-11-27 15:14:58.053543] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:30.140 [2024-11-27 15:14:58.053571] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:10:30.140 [2024-11-27 15:14:58.053581] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:30.140 [2024-11-27 15:14:58.053873] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:30.140 [2024-11-27 15:14:58.053891] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:30.140 [2024-11-27 15:14:58.053954] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:10:30.140 [2024-11-27 15:14:58.053974] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:30.140 [2024-11-27 15:14:58.054067] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:10:30.140 [2024-11-27 15:14:58.054079] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:30.140 [2024-11-27 15:14:58.054297] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:10:30.140 [2024-11-27 15:14:58.054412] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:10:30.140 [2024-11-27 15:14:58.054426] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:10:30.140 [2024-11-27 15:14:58.054520] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:30.140 pt4 00:10:30.140 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.140 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:30.140 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:30.140 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:30.140 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:30.140 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:30.140 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:30.140 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:30.140 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:30.140 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:30.140 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:30.140 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:30.140 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:30.140 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:30.140 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:30.140 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.140 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.140 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.140 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:30.140 "name": "raid_bdev1", 00:10:30.140 "uuid": "7efd1cef-d739-40ca-a820-d086958ecd0d", 00:10:30.140 "strip_size_kb": 64, 00:10:30.140 "state": "online", 00:10:30.140 "raid_level": "concat", 00:10:30.140 "superblock": true, 00:10:30.140 "num_base_bdevs": 4, 00:10:30.140 "num_base_bdevs_discovered": 4, 00:10:30.140 "num_base_bdevs_operational": 4, 00:10:30.140 "base_bdevs_list": [ 00:10:30.140 { 00:10:30.140 "name": "pt1", 00:10:30.140 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:30.140 "is_configured": true, 00:10:30.140 "data_offset": 2048, 00:10:30.140 "data_size": 63488 00:10:30.140 }, 00:10:30.140 { 00:10:30.140 "name": "pt2", 00:10:30.140 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:30.140 "is_configured": true, 00:10:30.140 "data_offset": 2048, 00:10:30.140 "data_size": 63488 00:10:30.140 }, 00:10:30.140 { 00:10:30.140 "name": "pt3", 00:10:30.140 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:30.140 "is_configured": true, 00:10:30.140 "data_offset": 2048, 00:10:30.140 "data_size": 63488 00:10:30.140 }, 00:10:30.140 { 00:10:30.140 "name": "pt4", 00:10:30.140 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:30.140 "is_configured": true, 00:10:30.140 "data_offset": 2048, 00:10:30.140 "data_size": 63488 00:10:30.140 } 00:10:30.140 ] 00:10:30.140 }' 00:10:30.140 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:30.140 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.415 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:10:30.415 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:30.415 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:30.415 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:30.415 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:30.415 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:30.415 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:30.415 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:30.415 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.415 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.415 [2024-11-27 15:14:58.469199] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:30.415 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.415 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:30.415 "name": "raid_bdev1", 00:10:30.415 "aliases": [ 00:10:30.415 "7efd1cef-d739-40ca-a820-d086958ecd0d" 00:10:30.415 ], 00:10:30.415 "product_name": "Raid Volume", 00:10:30.415 "block_size": 512, 00:10:30.415 "num_blocks": 253952, 00:10:30.415 "uuid": "7efd1cef-d739-40ca-a820-d086958ecd0d", 00:10:30.415 "assigned_rate_limits": { 00:10:30.415 "rw_ios_per_sec": 0, 00:10:30.415 "rw_mbytes_per_sec": 0, 00:10:30.415 "r_mbytes_per_sec": 0, 00:10:30.415 "w_mbytes_per_sec": 0 00:10:30.415 }, 00:10:30.415 "claimed": false, 00:10:30.415 "zoned": false, 00:10:30.415 "supported_io_types": { 00:10:30.415 "read": true, 00:10:30.415 "write": true, 00:10:30.415 "unmap": true, 00:10:30.415 "flush": true, 00:10:30.415 "reset": true, 00:10:30.415 "nvme_admin": false, 00:10:30.415 "nvme_io": false, 00:10:30.415 "nvme_io_md": false, 00:10:30.415 "write_zeroes": true, 00:10:30.415 "zcopy": false, 00:10:30.415 "get_zone_info": false, 00:10:30.415 "zone_management": false, 00:10:30.415 "zone_append": false, 00:10:30.415 "compare": false, 00:10:30.415 "compare_and_write": false, 00:10:30.415 "abort": false, 00:10:30.415 "seek_hole": false, 00:10:30.415 "seek_data": false, 00:10:30.415 "copy": false, 00:10:30.415 "nvme_iov_md": false 00:10:30.415 }, 00:10:30.415 "memory_domains": [ 00:10:30.415 { 00:10:30.415 "dma_device_id": "system", 00:10:30.415 "dma_device_type": 1 00:10:30.415 }, 00:10:30.415 { 00:10:30.415 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:30.415 "dma_device_type": 2 00:10:30.415 }, 00:10:30.415 { 00:10:30.415 "dma_device_id": "system", 00:10:30.415 "dma_device_type": 1 00:10:30.415 }, 00:10:30.415 { 00:10:30.415 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:30.415 "dma_device_type": 2 00:10:30.415 }, 00:10:30.415 { 00:10:30.415 "dma_device_id": "system", 00:10:30.415 "dma_device_type": 1 00:10:30.415 }, 00:10:30.415 { 00:10:30.415 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:30.415 "dma_device_type": 2 00:10:30.415 }, 00:10:30.415 { 00:10:30.415 "dma_device_id": "system", 00:10:30.415 "dma_device_type": 1 00:10:30.415 }, 00:10:30.415 { 00:10:30.415 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:30.415 "dma_device_type": 2 00:10:30.415 } 00:10:30.415 ], 00:10:30.415 "driver_specific": { 00:10:30.415 "raid": { 00:10:30.415 "uuid": "7efd1cef-d739-40ca-a820-d086958ecd0d", 00:10:30.415 "strip_size_kb": 64, 00:10:30.415 "state": "online", 00:10:30.415 "raid_level": "concat", 00:10:30.415 "superblock": true, 00:10:30.415 "num_base_bdevs": 4, 00:10:30.415 "num_base_bdevs_discovered": 4, 00:10:30.415 "num_base_bdevs_operational": 4, 00:10:30.415 "base_bdevs_list": [ 00:10:30.415 { 00:10:30.415 "name": "pt1", 00:10:30.415 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:30.415 "is_configured": true, 00:10:30.415 "data_offset": 2048, 00:10:30.415 "data_size": 63488 00:10:30.415 }, 00:10:30.415 { 00:10:30.415 "name": "pt2", 00:10:30.415 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:30.415 "is_configured": true, 00:10:30.415 "data_offset": 2048, 00:10:30.415 "data_size": 63488 00:10:30.415 }, 00:10:30.415 { 00:10:30.415 "name": "pt3", 00:10:30.415 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:30.415 "is_configured": true, 00:10:30.415 "data_offset": 2048, 00:10:30.415 "data_size": 63488 00:10:30.416 }, 00:10:30.416 { 00:10:30.416 "name": "pt4", 00:10:30.416 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:30.416 "is_configured": true, 00:10:30.416 "data_offset": 2048, 00:10:30.416 "data_size": 63488 00:10:30.416 } 00:10:30.416 ] 00:10:30.416 } 00:10:30.416 } 00:10:30.416 }' 00:10:30.416 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:30.676 pt2 00:10:30.676 pt3 00:10:30.676 pt4' 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.676 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.676 [2024-11-27 15:14:58.768728] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:30.936 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.936 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 7efd1cef-d739-40ca-a820-d086958ecd0d '!=' 7efd1cef-d739-40ca-a820-d086958ecd0d ']' 00:10:30.936 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:10:30.936 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:30.936 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:30.936 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 83617 00:10:30.936 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 83617 ']' 00:10:30.936 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 83617 00:10:30.936 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:10:30.936 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:30.936 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 83617 00:10:30.936 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:30.936 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:30.936 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 83617' 00:10:30.936 killing process with pid 83617 00:10:30.936 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 83617 00:10:30.936 [2024-11-27 15:14:58.844667] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:30.936 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 83617 00:10:30.936 [2024-11-27 15:14:58.844873] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:30.936 [2024-11-27 15:14:58.844958] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:30.936 [2024-11-27 15:14:58.844971] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:10:30.936 [2024-11-27 15:14:58.888680] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:31.195 15:14:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:10:31.195 00:10:31.195 real 0m4.063s 00:10:31.195 user 0m6.357s 00:10:31.195 sys 0m0.962s 00:10:31.195 15:14:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:31.195 15:14:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.195 ************************************ 00:10:31.195 END TEST raid_superblock_test 00:10:31.195 ************************************ 00:10:31.195 15:14:59 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 4 read 00:10:31.195 15:14:59 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:31.195 15:14:59 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:31.195 15:14:59 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:31.195 ************************************ 00:10:31.195 START TEST raid_read_error_test 00:10:31.195 ************************************ 00:10:31.195 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 4 read 00:10:31.195 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:10:31.195 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:10:31.195 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:10:31.195 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:31.195 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:31.195 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:31.195 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:31.195 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:31.195 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:31.195 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:31.195 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:31.195 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:31.195 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:31.195 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:31.195 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:10:31.195 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:31.195 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:31.196 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:31.196 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:31.196 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:31.196 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:31.196 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:31.196 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:31.196 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:31.196 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:10:31.196 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:10:31.196 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:10:31.196 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:31.196 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.cRjSmj1ZWV 00:10:31.196 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=83861 00:10:31.196 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:31.196 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 83861 00:10:31.196 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 83861 ']' 00:10:31.196 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:31.196 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:31.196 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:31.196 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:31.196 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:31.196 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.196 [2024-11-27 15:14:59.297095] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:10:31.196 [2024-11-27 15:14:59.297423] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid83861 ] 00:10:31.455 [2024-11-27 15:14:59.474445] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:31.455 [2024-11-27 15:14:59.500108] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:31.455 [2024-11-27 15:14:59.543143] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:31.455 [2024-11-27 15:14:59.543252] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:32.392 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:32.392 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.393 BaseBdev1_malloc 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.393 true 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.393 [2024-11-27 15:15:00.155384] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:32.393 [2024-11-27 15:15:00.155498] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:32.393 [2024-11-27 15:15:00.155546] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:10:32.393 [2024-11-27 15:15:00.155602] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:32.393 [2024-11-27 15:15:00.157775] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:32.393 [2024-11-27 15:15:00.157848] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:32.393 BaseBdev1 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.393 BaseBdev2_malloc 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.393 true 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.393 [2024-11-27 15:15:00.184015] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:32.393 [2024-11-27 15:15:00.184104] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:32.393 [2024-11-27 15:15:00.184125] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:10:32.393 [2024-11-27 15:15:00.184134] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:32.393 [2024-11-27 15:15:00.186175] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:32.393 [2024-11-27 15:15:00.186210] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:32.393 BaseBdev2 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.393 BaseBdev3_malloc 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.393 true 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.393 [2024-11-27 15:15:00.212589] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:32.393 [2024-11-27 15:15:00.212634] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:32.393 [2024-11-27 15:15:00.212652] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:10:32.393 [2024-11-27 15:15:00.212660] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:32.393 [2024-11-27 15:15:00.214688] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:32.393 [2024-11-27 15:15:00.214724] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:32.393 BaseBdev3 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.393 BaseBdev4_malloc 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.393 true 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.393 [2024-11-27 15:15:00.252968] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:10:32.393 [2024-11-27 15:15:00.253016] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:32.393 [2024-11-27 15:15:00.253039] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:10:32.393 [2024-11-27 15:15:00.253049] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:32.393 [2024-11-27 15:15:00.255075] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:32.393 [2024-11-27 15:15:00.255176] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:10:32.393 BaseBdev4 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.393 [2024-11-27 15:15:00.261003] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:32.393 [2024-11-27 15:15:00.262747] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:32.393 [2024-11-27 15:15:00.262888] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:32.393 [2024-11-27 15:15:00.262960] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:32.393 [2024-11-27 15:15:00.263154] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007080 00:10:32.393 [2024-11-27 15:15:00.263171] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:32.393 [2024-11-27 15:15:00.263424] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006700 00:10:32.393 [2024-11-27 15:15:00.263551] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007080 00:10:32.393 [2024-11-27 15:15:00.263573] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007080 00:10:32.393 [2024-11-27 15:15:00.263720] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.393 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:32.394 "name": "raid_bdev1", 00:10:32.394 "uuid": "8d595e45-7f81-4b0f-84ec-443ba371874d", 00:10:32.394 "strip_size_kb": 64, 00:10:32.394 "state": "online", 00:10:32.394 "raid_level": "concat", 00:10:32.394 "superblock": true, 00:10:32.394 "num_base_bdevs": 4, 00:10:32.394 "num_base_bdevs_discovered": 4, 00:10:32.394 "num_base_bdevs_operational": 4, 00:10:32.394 "base_bdevs_list": [ 00:10:32.394 { 00:10:32.394 "name": "BaseBdev1", 00:10:32.394 "uuid": "cde5dc46-2a00-566d-8756-6fab8d462cf5", 00:10:32.394 "is_configured": true, 00:10:32.394 "data_offset": 2048, 00:10:32.394 "data_size": 63488 00:10:32.394 }, 00:10:32.394 { 00:10:32.394 "name": "BaseBdev2", 00:10:32.394 "uuid": "10bbd0cb-5856-58ed-bd8e-7c9441f71d42", 00:10:32.394 "is_configured": true, 00:10:32.394 "data_offset": 2048, 00:10:32.394 "data_size": 63488 00:10:32.394 }, 00:10:32.394 { 00:10:32.394 "name": "BaseBdev3", 00:10:32.394 "uuid": "43f47ece-2c47-5509-bf5d-e92a64df86d9", 00:10:32.394 "is_configured": true, 00:10:32.394 "data_offset": 2048, 00:10:32.394 "data_size": 63488 00:10:32.394 }, 00:10:32.394 { 00:10:32.394 "name": "BaseBdev4", 00:10:32.394 "uuid": "017e6881-cf33-50f1-9e36-5249f6ab679c", 00:10:32.394 "is_configured": true, 00:10:32.394 "data_offset": 2048, 00:10:32.394 "data_size": 63488 00:10:32.394 } 00:10:32.394 ] 00:10:32.394 }' 00:10:32.394 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:32.394 15:15:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.653 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:32.653 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:32.912 [2024-11-27 15:15:00.784482] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000068a0 00:10:33.850 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:10:33.850 15:15:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.850 15:15:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.850 15:15:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.850 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:33.850 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:10:33.850 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:10:33.850 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:33.850 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:33.850 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:33.850 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:33.850 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:33.850 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:33.850 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:33.850 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:33.850 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:33.850 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:33.851 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:33.851 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:33.851 15:15:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.851 15:15:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.851 15:15:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.851 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:33.851 "name": "raid_bdev1", 00:10:33.851 "uuid": "8d595e45-7f81-4b0f-84ec-443ba371874d", 00:10:33.851 "strip_size_kb": 64, 00:10:33.851 "state": "online", 00:10:33.851 "raid_level": "concat", 00:10:33.851 "superblock": true, 00:10:33.851 "num_base_bdevs": 4, 00:10:33.851 "num_base_bdevs_discovered": 4, 00:10:33.851 "num_base_bdevs_operational": 4, 00:10:33.851 "base_bdevs_list": [ 00:10:33.851 { 00:10:33.851 "name": "BaseBdev1", 00:10:33.851 "uuid": "cde5dc46-2a00-566d-8756-6fab8d462cf5", 00:10:33.851 "is_configured": true, 00:10:33.851 "data_offset": 2048, 00:10:33.851 "data_size": 63488 00:10:33.851 }, 00:10:33.851 { 00:10:33.851 "name": "BaseBdev2", 00:10:33.851 "uuid": "10bbd0cb-5856-58ed-bd8e-7c9441f71d42", 00:10:33.851 "is_configured": true, 00:10:33.851 "data_offset": 2048, 00:10:33.851 "data_size": 63488 00:10:33.851 }, 00:10:33.851 { 00:10:33.851 "name": "BaseBdev3", 00:10:33.851 "uuid": "43f47ece-2c47-5509-bf5d-e92a64df86d9", 00:10:33.851 "is_configured": true, 00:10:33.851 "data_offset": 2048, 00:10:33.851 "data_size": 63488 00:10:33.851 }, 00:10:33.851 { 00:10:33.851 "name": "BaseBdev4", 00:10:33.851 "uuid": "017e6881-cf33-50f1-9e36-5249f6ab679c", 00:10:33.851 "is_configured": true, 00:10:33.851 "data_offset": 2048, 00:10:33.851 "data_size": 63488 00:10:33.851 } 00:10:33.851 ] 00:10:33.851 }' 00:10:33.851 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:33.851 15:15:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.111 15:15:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:34.111 15:15:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.111 15:15:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.111 [2024-11-27 15:15:02.083890] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:34.111 [2024-11-27 15:15:02.084023] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:34.111 [2024-11-27 15:15:02.086707] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:34.111 [2024-11-27 15:15:02.086811] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:34.111 [2024-11-27 15:15:02.086893] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:34.111 [2024-11-27 15:15:02.086960] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007080 name raid_bdev1, state offline 00:10:34.111 { 00:10:34.111 "results": [ 00:10:34.111 { 00:10:34.111 "job": "raid_bdev1", 00:10:34.111 "core_mask": "0x1", 00:10:34.111 "workload": "randrw", 00:10:34.111 "percentage": 50, 00:10:34.111 "status": "finished", 00:10:34.111 "queue_depth": 1, 00:10:34.111 "io_size": 131072, 00:10:34.111 "runtime": 1.300183, 00:10:34.111 "iops": 16017.745194330337, 00:10:34.111 "mibps": 2002.218149291292, 00:10:34.111 "io_failed": 1, 00:10:34.111 "io_timeout": 0, 00:10:34.111 "avg_latency_us": 86.39645136488305, 00:10:34.111 "min_latency_us": 26.270742358078603, 00:10:34.111 "max_latency_us": 1409.4532751091704 00:10:34.111 } 00:10:34.111 ], 00:10:34.111 "core_count": 1 00:10:34.111 } 00:10:34.111 15:15:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.111 15:15:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 83861 00:10:34.111 15:15:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 83861 ']' 00:10:34.111 15:15:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 83861 00:10:34.111 15:15:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:10:34.111 15:15:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:34.111 15:15:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 83861 00:10:34.111 killing process with pid 83861 00:10:34.111 15:15:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:34.111 15:15:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:34.111 15:15:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 83861' 00:10:34.111 15:15:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 83861 00:10:34.111 [2024-11-27 15:15:02.135260] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:34.111 15:15:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 83861 00:10:34.111 [2024-11-27 15:15:02.170612] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:34.371 15:15:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.cRjSmj1ZWV 00:10:34.371 15:15:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:34.371 15:15:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:34.371 ************************************ 00:10:34.371 END TEST raid_read_error_test 00:10:34.371 ************************************ 00:10:34.371 15:15:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.77 00:10:34.371 15:15:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:10:34.371 15:15:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:34.371 15:15:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:34.371 15:15:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.77 != \0\.\0\0 ]] 00:10:34.371 00:10:34.371 real 0m3.206s 00:10:34.371 user 0m3.992s 00:10:34.371 sys 0m0.543s 00:10:34.371 15:15:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:34.371 15:15:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.371 15:15:02 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 4 write 00:10:34.371 15:15:02 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:34.371 15:15:02 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:34.371 15:15:02 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:34.371 ************************************ 00:10:34.371 START TEST raid_write_error_test 00:10:34.371 ************************************ 00:10:34.371 15:15:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 4 write 00:10:34.371 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:10:34.371 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:10:34.371 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:10:34.371 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:34.371 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:34.371 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:34.371 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:34.371 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:34.371 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:34.371 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:34.371 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:34.371 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:34.371 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:34.371 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:34.371 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:10:34.371 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:34.371 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:34.371 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:34.371 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:34.371 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:34.371 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:34.371 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:34.371 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:34.371 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:34.371 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:10:34.371 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:10:34.371 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:10:34.630 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:34.630 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.yz2tWVP6cQ 00:10:34.630 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=83990 00:10:34.630 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:34.630 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 83990 00:10:34.630 15:15:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 83990 ']' 00:10:34.630 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:34.630 15:15:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:34.630 15:15:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:34.630 15:15:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:34.630 15:15:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:34.630 15:15:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.630 [2024-11-27 15:15:02.577254] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:10:34.630 [2024-11-27 15:15:02.577398] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid83990 ] 00:10:34.888 [2024-11-27 15:15:02.752824] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:34.888 [2024-11-27 15:15:02.778308] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:34.888 [2024-11-27 15:15:02.821188] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:34.888 [2024-11-27 15:15:02.821226] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.454 BaseBdev1_malloc 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.454 true 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.454 [2024-11-27 15:15:03.417569] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:35.454 [2024-11-27 15:15:03.417640] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:35.454 [2024-11-27 15:15:03.417689] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:10:35.454 [2024-11-27 15:15:03.417700] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:35.454 [2024-11-27 15:15:03.419845] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:35.454 [2024-11-27 15:15:03.419967] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:35.454 BaseBdev1 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.454 BaseBdev2_malloc 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.454 true 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.454 [2024-11-27 15:15:03.458260] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:35.454 [2024-11-27 15:15:03.458341] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:35.454 [2024-11-27 15:15:03.458365] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:10:35.454 [2024-11-27 15:15:03.458375] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:35.454 [2024-11-27 15:15:03.460703] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:35.454 [2024-11-27 15:15:03.460856] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:35.454 BaseBdev2 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.454 BaseBdev3_malloc 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.454 true 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.454 [2024-11-27 15:15:03.495319] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:35.454 [2024-11-27 15:15:03.495481] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:35.454 [2024-11-27 15:15:03.495510] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:10:35.454 [2024-11-27 15:15:03.495519] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:35.454 [2024-11-27 15:15:03.497728] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:35.454 [2024-11-27 15:15:03.497770] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:35.454 BaseBdev3 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.454 BaseBdev4_malloc 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.454 true 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.454 [2024-11-27 15:15:03.539705] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:10:35.454 [2024-11-27 15:15:03.539774] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:35.454 [2024-11-27 15:15:03.539801] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:10:35.454 [2024-11-27 15:15:03.539811] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:35.454 [2024-11-27 15:15:03.542029] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:35.454 [2024-11-27 15:15:03.542070] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:10:35.454 BaseBdev4 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.454 [2024-11-27 15:15:03.547765] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:35.454 [2024-11-27 15:15:03.549717] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:35.454 [2024-11-27 15:15:03.549886] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:35.454 [2024-11-27 15:15:03.549965] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:35.454 [2024-11-27 15:15:03.550183] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007080 00:10:35.454 [2024-11-27 15:15:03.550196] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:35.454 [2024-11-27 15:15:03.550505] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006700 00:10:35.454 [2024-11-27 15:15:03.550658] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007080 00:10:35.454 [2024-11-27 15:15:03.550671] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007080 00:10:35.454 [2024-11-27 15:15:03.550847] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:35.454 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:35.714 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.714 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:35.714 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.714 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.714 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.714 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:35.714 "name": "raid_bdev1", 00:10:35.715 "uuid": "3f7ec965-9579-482a-8411-8a999fda8674", 00:10:35.715 "strip_size_kb": 64, 00:10:35.715 "state": "online", 00:10:35.715 "raid_level": "concat", 00:10:35.715 "superblock": true, 00:10:35.715 "num_base_bdevs": 4, 00:10:35.715 "num_base_bdevs_discovered": 4, 00:10:35.715 "num_base_bdevs_operational": 4, 00:10:35.715 "base_bdevs_list": [ 00:10:35.715 { 00:10:35.715 "name": "BaseBdev1", 00:10:35.715 "uuid": "9f00ffba-5cb5-5b79-88e4-33231c7289b6", 00:10:35.715 "is_configured": true, 00:10:35.715 "data_offset": 2048, 00:10:35.715 "data_size": 63488 00:10:35.715 }, 00:10:35.715 { 00:10:35.715 "name": "BaseBdev2", 00:10:35.715 "uuid": "a6d99a76-fbc3-52ce-8986-88ccda08b928", 00:10:35.715 "is_configured": true, 00:10:35.715 "data_offset": 2048, 00:10:35.715 "data_size": 63488 00:10:35.715 }, 00:10:35.715 { 00:10:35.715 "name": "BaseBdev3", 00:10:35.715 "uuid": "b5be2458-c4d2-50c7-a7d4-af0ba689f56b", 00:10:35.715 "is_configured": true, 00:10:35.715 "data_offset": 2048, 00:10:35.715 "data_size": 63488 00:10:35.715 }, 00:10:35.715 { 00:10:35.715 "name": "BaseBdev4", 00:10:35.715 "uuid": "f61f5100-5565-5079-8d14-4859ea0cc40c", 00:10:35.715 "is_configured": true, 00:10:35.715 "data_offset": 2048, 00:10:35.715 "data_size": 63488 00:10:35.715 } 00:10:35.715 ] 00:10:35.715 }' 00:10:35.715 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:35.715 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.974 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:35.974 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:35.974 [2024-11-27 15:15:04.043262] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000068a0 00:10:36.914 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:10:36.914 15:15:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.914 15:15:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.914 15:15:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.914 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:36.914 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:10:36.914 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:10:36.914 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:36.914 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:36.914 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:36.914 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:36.914 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:36.914 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:36.914 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:36.914 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:36.914 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:36.914 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:36.914 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:36.914 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:36.914 15:15:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.914 15:15:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.914 15:15:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.914 15:15:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:36.914 "name": "raid_bdev1", 00:10:36.914 "uuid": "3f7ec965-9579-482a-8411-8a999fda8674", 00:10:36.914 "strip_size_kb": 64, 00:10:36.914 "state": "online", 00:10:36.914 "raid_level": "concat", 00:10:36.914 "superblock": true, 00:10:36.914 "num_base_bdevs": 4, 00:10:36.914 "num_base_bdevs_discovered": 4, 00:10:36.914 "num_base_bdevs_operational": 4, 00:10:36.914 "base_bdevs_list": [ 00:10:36.914 { 00:10:36.914 "name": "BaseBdev1", 00:10:36.914 "uuid": "9f00ffba-5cb5-5b79-88e4-33231c7289b6", 00:10:36.914 "is_configured": true, 00:10:36.914 "data_offset": 2048, 00:10:36.914 "data_size": 63488 00:10:36.914 }, 00:10:36.914 { 00:10:36.915 "name": "BaseBdev2", 00:10:36.915 "uuid": "a6d99a76-fbc3-52ce-8986-88ccda08b928", 00:10:36.915 "is_configured": true, 00:10:36.915 "data_offset": 2048, 00:10:36.915 "data_size": 63488 00:10:36.915 }, 00:10:36.915 { 00:10:36.915 "name": "BaseBdev3", 00:10:36.915 "uuid": "b5be2458-c4d2-50c7-a7d4-af0ba689f56b", 00:10:36.915 "is_configured": true, 00:10:36.915 "data_offset": 2048, 00:10:36.915 "data_size": 63488 00:10:36.915 }, 00:10:36.915 { 00:10:36.915 "name": "BaseBdev4", 00:10:36.915 "uuid": "f61f5100-5565-5079-8d14-4859ea0cc40c", 00:10:36.915 "is_configured": true, 00:10:36.915 "data_offset": 2048, 00:10:36.915 "data_size": 63488 00:10:36.915 } 00:10:36.915 ] 00:10:36.915 }' 00:10:36.915 15:15:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:36.915 15:15:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.485 15:15:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:37.485 15:15:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.485 15:15:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.485 [2024-11-27 15:15:05.447468] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:37.485 [2024-11-27 15:15:05.447633] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:37.485 [2024-11-27 15:15:05.450312] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:37.485 [2024-11-27 15:15:05.450365] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:37.485 [2024-11-27 15:15:05.450407] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:37.485 [2024-11-27 15:15:05.450417] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007080 name raid_bdev1, state offline 00:10:37.485 { 00:10:37.485 "results": [ 00:10:37.485 { 00:10:37.485 "job": "raid_bdev1", 00:10:37.485 "core_mask": "0x1", 00:10:37.485 "workload": "randrw", 00:10:37.485 "percentage": 50, 00:10:37.485 "status": "finished", 00:10:37.485 "queue_depth": 1, 00:10:37.485 "io_size": 131072, 00:10:37.485 "runtime": 1.405198, 00:10:37.485 "iops": 16259.63031544309, 00:10:37.485 "mibps": 2032.4537894303862, 00:10:37.485 "io_failed": 1, 00:10:37.485 "io_timeout": 0, 00:10:37.485 "avg_latency_us": 84.93710915081184, 00:10:37.485 "min_latency_us": 24.817467248908297, 00:10:37.485 "max_latency_us": 1366.5257641921398 00:10:37.485 } 00:10:37.485 ], 00:10:37.485 "core_count": 1 00:10:37.485 } 00:10:37.485 15:15:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.485 15:15:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 83990 00:10:37.485 15:15:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 83990 ']' 00:10:37.485 15:15:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 83990 00:10:37.485 15:15:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:10:37.485 15:15:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:37.485 15:15:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 83990 00:10:37.485 killing process with pid 83990 00:10:37.485 15:15:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:37.485 15:15:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:37.485 15:15:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 83990' 00:10:37.485 15:15:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 83990 00:10:37.485 [2024-11-27 15:15:05.497371] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:37.485 15:15:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 83990 00:10:37.485 [2024-11-27 15:15:05.532181] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:37.744 15:15:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.yz2tWVP6cQ 00:10:37.744 15:15:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:37.744 15:15:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:37.744 15:15:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.71 00:10:37.744 15:15:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:10:37.744 15:15:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:37.744 15:15:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:37.744 15:15:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.71 != \0\.\0\0 ]] 00:10:37.744 00:10:37.744 real 0m3.296s 00:10:37.744 user 0m4.136s 00:10:37.744 sys 0m0.569s 00:10:37.744 ************************************ 00:10:37.744 END TEST raid_write_error_test 00:10:37.744 ************************************ 00:10:37.744 15:15:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:37.744 15:15:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.744 15:15:05 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:10:37.744 15:15:05 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 4 false 00:10:37.745 15:15:05 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:37.745 15:15:05 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:37.745 15:15:05 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:37.745 ************************************ 00:10:37.745 START TEST raid_state_function_test 00:10:37.745 ************************************ 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 4 false 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:10:37.745 Process raid pid: 84125 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=84125 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 84125' 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 84125 00:10:37.745 15:15:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 84125 ']' 00:10:38.004 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:38.004 15:15:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:38.004 15:15:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:38.004 15:15:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:38.004 15:15:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:38.004 15:15:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.004 [2024-11-27 15:15:05.928872] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:10:38.004 [2024-11-27 15:15:05.929010] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:38.004 [2024-11-27 15:15:06.099917] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:38.264 [2024-11-27 15:15:06.126491] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:38.264 [2024-11-27 15:15:06.169429] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:38.264 [2024-11-27 15:15:06.169471] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:38.833 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:38.833 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:10:38.833 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:38.833 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.833 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.833 [2024-11-27 15:15:06.752547] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:38.833 [2024-11-27 15:15:06.752614] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:38.833 [2024-11-27 15:15:06.752625] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:38.833 [2024-11-27 15:15:06.752636] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:38.833 [2024-11-27 15:15:06.752642] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:38.833 [2024-11-27 15:15:06.752653] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:38.833 [2024-11-27 15:15:06.752661] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:38.833 [2024-11-27 15:15:06.752670] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:38.833 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.833 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:38.833 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:38.833 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:38.833 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:38.833 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:38.833 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:38.833 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:38.834 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:38.834 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:38.834 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:38.834 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:38.834 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:38.834 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.834 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.834 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.834 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:38.834 "name": "Existed_Raid", 00:10:38.834 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:38.834 "strip_size_kb": 0, 00:10:38.834 "state": "configuring", 00:10:38.834 "raid_level": "raid1", 00:10:38.834 "superblock": false, 00:10:38.834 "num_base_bdevs": 4, 00:10:38.834 "num_base_bdevs_discovered": 0, 00:10:38.834 "num_base_bdevs_operational": 4, 00:10:38.834 "base_bdevs_list": [ 00:10:38.834 { 00:10:38.834 "name": "BaseBdev1", 00:10:38.834 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:38.834 "is_configured": false, 00:10:38.834 "data_offset": 0, 00:10:38.834 "data_size": 0 00:10:38.834 }, 00:10:38.834 { 00:10:38.834 "name": "BaseBdev2", 00:10:38.834 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:38.834 "is_configured": false, 00:10:38.834 "data_offset": 0, 00:10:38.834 "data_size": 0 00:10:38.834 }, 00:10:38.834 { 00:10:38.834 "name": "BaseBdev3", 00:10:38.834 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:38.834 "is_configured": false, 00:10:38.834 "data_offset": 0, 00:10:38.834 "data_size": 0 00:10:38.834 }, 00:10:38.834 { 00:10:38.834 "name": "BaseBdev4", 00:10:38.834 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:38.834 "is_configured": false, 00:10:38.834 "data_offset": 0, 00:10:38.834 "data_size": 0 00:10:38.834 } 00:10:38.834 ] 00:10:38.834 }' 00:10:38.834 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:38.834 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.402 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:39.402 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:39.402 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.402 [2024-11-27 15:15:07.235685] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:39.402 [2024-11-27 15:15:07.235725] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:10:39.402 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.402 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:39.402 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:39.402 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.402 [2024-11-27 15:15:07.247699] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:39.402 [2024-11-27 15:15:07.247743] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:39.402 [2024-11-27 15:15:07.247752] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:39.402 [2024-11-27 15:15:07.247761] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:39.402 [2024-11-27 15:15:07.247767] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:39.402 [2024-11-27 15:15:07.247776] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:39.402 [2024-11-27 15:15:07.247782] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:39.402 [2024-11-27 15:15:07.247791] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:39.402 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.402 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:39.402 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:39.402 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.402 [2024-11-27 15:15:07.268552] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:39.402 BaseBdev1 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.403 [ 00:10:39.403 { 00:10:39.403 "name": "BaseBdev1", 00:10:39.403 "aliases": [ 00:10:39.403 "b50ed93b-bda1-45de-92c7-9eeb2ca10585" 00:10:39.403 ], 00:10:39.403 "product_name": "Malloc disk", 00:10:39.403 "block_size": 512, 00:10:39.403 "num_blocks": 65536, 00:10:39.403 "uuid": "b50ed93b-bda1-45de-92c7-9eeb2ca10585", 00:10:39.403 "assigned_rate_limits": { 00:10:39.403 "rw_ios_per_sec": 0, 00:10:39.403 "rw_mbytes_per_sec": 0, 00:10:39.403 "r_mbytes_per_sec": 0, 00:10:39.403 "w_mbytes_per_sec": 0 00:10:39.403 }, 00:10:39.403 "claimed": true, 00:10:39.403 "claim_type": "exclusive_write", 00:10:39.403 "zoned": false, 00:10:39.403 "supported_io_types": { 00:10:39.403 "read": true, 00:10:39.403 "write": true, 00:10:39.403 "unmap": true, 00:10:39.403 "flush": true, 00:10:39.403 "reset": true, 00:10:39.403 "nvme_admin": false, 00:10:39.403 "nvme_io": false, 00:10:39.403 "nvme_io_md": false, 00:10:39.403 "write_zeroes": true, 00:10:39.403 "zcopy": true, 00:10:39.403 "get_zone_info": false, 00:10:39.403 "zone_management": false, 00:10:39.403 "zone_append": false, 00:10:39.403 "compare": false, 00:10:39.403 "compare_and_write": false, 00:10:39.403 "abort": true, 00:10:39.403 "seek_hole": false, 00:10:39.403 "seek_data": false, 00:10:39.403 "copy": true, 00:10:39.403 "nvme_iov_md": false 00:10:39.403 }, 00:10:39.403 "memory_domains": [ 00:10:39.403 { 00:10:39.403 "dma_device_id": "system", 00:10:39.403 "dma_device_type": 1 00:10:39.403 }, 00:10:39.403 { 00:10:39.403 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:39.403 "dma_device_type": 2 00:10:39.403 } 00:10:39.403 ], 00:10:39.403 "driver_specific": {} 00:10:39.403 } 00:10:39.403 ] 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:39.403 "name": "Existed_Raid", 00:10:39.403 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:39.403 "strip_size_kb": 0, 00:10:39.403 "state": "configuring", 00:10:39.403 "raid_level": "raid1", 00:10:39.403 "superblock": false, 00:10:39.403 "num_base_bdevs": 4, 00:10:39.403 "num_base_bdevs_discovered": 1, 00:10:39.403 "num_base_bdevs_operational": 4, 00:10:39.403 "base_bdevs_list": [ 00:10:39.403 { 00:10:39.403 "name": "BaseBdev1", 00:10:39.403 "uuid": "b50ed93b-bda1-45de-92c7-9eeb2ca10585", 00:10:39.403 "is_configured": true, 00:10:39.403 "data_offset": 0, 00:10:39.403 "data_size": 65536 00:10:39.403 }, 00:10:39.403 { 00:10:39.403 "name": "BaseBdev2", 00:10:39.403 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:39.403 "is_configured": false, 00:10:39.403 "data_offset": 0, 00:10:39.403 "data_size": 0 00:10:39.403 }, 00:10:39.403 { 00:10:39.403 "name": "BaseBdev3", 00:10:39.403 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:39.403 "is_configured": false, 00:10:39.403 "data_offset": 0, 00:10:39.403 "data_size": 0 00:10:39.403 }, 00:10:39.403 { 00:10:39.403 "name": "BaseBdev4", 00:10:39.403 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:39.403 "is_configured": false, 00:10:39.403 "data_offset": 0, 00:10:39.403 "data_size": 0 00:10:39.403 } 00:10:39.403 ] 00:10:39.403 }' 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:39.403 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.662 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:39.662 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:39.662 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.662 [2024-11-27 15:15:07.739816] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:39.662 [2024-11-27 15:15:07.739986] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:10:39.662 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.662 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:39.662 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:39.662 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.662 [2024-11-27 15:15:07.751792] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:39.662 [2024-11-27 15:15:07.753656] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:39.662 [2024-11-27 15:15:07.753701] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:39.662 [2024-11-27 15:15:07.753711] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:39.662 [2024-11-27 15:15:07.753719] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:39.662 [2024-11-27 15:15:07.753725] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:39.662 [2024-11-27 15:15:07.753733] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:39.662 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.662 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:10:39.662 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:39.662 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:39.662 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:39.662 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:39.663 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:39.663 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:39.663 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:39.663 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:39.663 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:39.663 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:39.663 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:39.663 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:39.663 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:39.663 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:39.663 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.922 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.922 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:39.922 "name": "Existed_Raid", 00:10:39.922 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:39.922 "strip_size_kb": 0, 00:10:39.922 "state": "configuring", 00:10:39.922 "raid_level": "raid1", 00:10:39.922 "superblock": false, 00:10:39.922 "num_base_bdevs": 4, 00:10:39.922 "num_base_bdevs_discovered": 1, 00:10:39.922 "num_base_bdevs_operational": 4, 00:10:39.922 "base_bdevs_list": [ 00:10:39.922 { 00:10:39.922 "name": "BaseBdev1", 00:10:39.922 "uuid": "b50ed93b-bda1-45de-92c7-9eeb2ca10585", 00:10:39.922 "is_configured": true, 00:10:39.922 "data_offset": 0, 00:10:39.922 "data_size": 65536 00:10:39.922 }, 00:10:39.922 { 00:10:39.922 "name": "BaseBdev2", 00:10:39.922 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:39.922 "is_configured": false, 00:10:39.922 "data_offset": 0, 00:10:39.922 "data_size": 0 00:10:39.922 }, 00:10:39.922 { 00:10:39.922 "name": "BaseBdev3", 00:10:39.922 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:39.922 "is_configured": false, 00:10:39.922 "data_offset": 0, 00:10:39.922 "data_size": 0 00:10:39.922 }, 00:10:39.922 { 00:10:39.922 "name": "BaseBdev4", 00:10:39.922 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:39.922 "is_configured": false, 00:10:39.922 "data_offset": 0, 00:10:39.922 "data_size": 0 00:10:39.922 } 00:10:39.922 ] 00:10:39.922 }' 00:10:39.922 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:39.922 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.181 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:40.181 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.181 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.181 [2024-11-27 15:15:08.182041] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:40.181 BaseBdev2 00:10:40.181 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.181 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:10:40.181 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:40.181 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:40.181 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:40.181 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:40.181 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:40.181 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:40.181 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.181 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.181 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.181 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:40.181 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.181 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.181 [ 00:10:40.181 { 00:10:40.181 "name": "BaseBdev2", 00:10:40.181 "aliases": [ 00:10:40.181 "f79959fd-9826-4204-82e0-64d6d4df0881" 00:10:40.181 ], 00:10:40.181 "product_name": "Malloc disk", 00:10:40.181 "block_size": 512, 00:10:40.181 "num_blocks": 65536, 00:10:40.181 "uuid": "f79959fd-9826-4204-82e0-64d6d4df0881", 00:10:40.182 "assigned_rate_limits": { 00:10:40.182 "rw_ios_per_sec": 0, 00:10:40.182 "rw_mbytes_per_sec": 0, 00:10:40.182 "r_mbytes_per_sec": 0, 00:10:40.182 "w_mbytes_per_sec": 0 00:10:40.182 }, 00:10:40.182 "claimed": true, 00:10:40.182 "claim_type": "exclusive_write", 00:10:40.182 "zoned": false, 00:10:40.182 "supported_io_types": { 00:10:40.182 "read": true, 00:10:40.182 "write": true, 00:10:40.182 "unmap": true, 00:10:40.182 "flush": true, 00:10:40.182 "reset": true, 00:10:40.182 "nvme_admin": false, 00:10:40.182 "nvme_io": false, 00:10:40.182 "nvme_io_md": false, 00:10:40.182 "write_zeroes": true, 00:10:40.182 "zcopy": true, 00:10:40.182 "get_zone_info": false, 00:10:40.182 "zone_management": false, 00:10:40.182 "zone_append": false, 00:10:40.182 "compare": false, 00:10:40.182 "compare_and_write": false, 00:10:40.182 "abort": true, 00:10:40.182 "seek_hole": false, 00:10:40.182 "seek_data": false, 00:10:40.182 "copy": true, 00:10:40.182 "nvme_iov_md": false 00:10:40.182 }, 00:10:40.182 "memory_domains": [ 00:10:40.182 { 00:10:40.182 "dma_device_id": "system", 00:10:40.182 "dma_device_type": 1 00:10:40.182 }, 00:10:40.182 { 00:10:40.182 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:40.182 "dma_device_type": 2 00:10:40.182 } 00:10:40.182 ], 00:10:40.182 "driver_specific": {} 00:10:40.182 } 00:10:40.182 ] 00:10:40.182 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.182 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:40.182 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:40.182 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:40.182 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:40.182 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:40.182 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:40.182 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:40.182 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:40.182 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:40.182 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:40.182 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:40.182 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:40.182 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:40.182 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:40.182 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:40.182 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.182 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.182 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.182 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:40.182 "name": "Existed_Raid", 00:10:40.182 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:40.182 "strip_size_kb": 0, 00:10:40.182 "state": "configuring", 00:10:40.182 "raid_level": "raid1", 00:10:40.182 "superblock": false, 00:10:40.182 "num_base_bdevs": 4, 00:10:40.182 "num_base_bdevs_discovered": 2, 00:10:40.182 "num_base_bdevs_operational": 4, 00:10:40.182 "base_bdevs_list": [ 00:10:40.182 { 00:10:40.182 "name": "BaseBdev1", 00:10:40.182 "uuid": "b50ed93b-bda1-45de-92c7-9eeb2ca10585", 00:10:40.182 "is_configured": true, 00:10:40.182 "data_offset": 0, 00:10:40.182 "data_size": 65536 00:10:40.182 }, 00:10:40.182 { 00:10:40.182 "name": "BaseBdev2", 00:10:40.182 "uuid": "f79959fd-9826-4204-82e0-64d6d4df0881", 00:10:40.182 "is_configured": true, 00:10:40.182 "data_offset": 0, 00:10:40.182 "data_size": 65536 00:10:40.182 }, 00:10:40.182 { 00:10:40.182 "name": "BaseBdev3", 00:10:40.182 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:40.182 "is_configured": false, 00:10:40.182 "data_offset": 0, 00:10:40.182 "data_size": 0 00:10:40.182 }, 00:10:40.182 { 00:10:40.182 "name": "BaseBdev4", 00:10:40.182 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:40.182 "is_configured": false, 00:10:40.182 "data_offset": 0, 00:10:40.182 "data_size": 0 00:10:40.182 } 00:10:40.182 ] 00:10:40.182 }' 00:10:40.182 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:40.182 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.751 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:40.751 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.751 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.751 [2024-11-27 15:15:08.674046] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:40.751 BaseBdev3 00:10:40.751 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.751 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:10:40.751 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:40.751 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:40.751 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:40.751 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:40.751 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:40.751 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:40.751 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.751 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.751 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.751 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:40.751 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.751 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.751 [ 00:10:40.751 { 00:10:40.751 "name": "BaseBdev3", 00:10:40.751 "aliases": [ 00:10:40.751 "c28b01c2-0a8c-42f5-becb-386c5ccfd2ec" 00:10:40.751 ], 00:10:40.751 "product_name": "Malloc disk", 00:10:40.751 "block_size": 512, 00:10:40.751 "num_blocks": 65536, 00:10:40.751 "uuid": "c28b01c2-0a8c-42f5-becb-386c5ccfd2ec", 00:10:40.751 "assigned_rate_limits": { 00:10:40.751 "rw_ios_per_sec": 0, 00:10:40.751 "rw_mbytes_per_sec": 0, 00:10:40.751 "r_mbytes_per_sec": 0, 00:10:40.751 "w_mbytes_per_sec": 0 00:10:40.751 }, 00:10:40.751 "claimed": true, 00:10:40.751 "claim_type": "exclusive_write", 00:10:40.751 "zoned": false, 00:10:40.751 "supported_io_types": { 00:10:40.752 "read": true, 00:10:40.752 "write": true, 00:10:40.752 "unmap": true, 00:10:40.752 "flush": true, 00:10:40.752 "reset": true, 00:10:40.752 "nvme_admin": false, 00:10:40.752 "nvme_io": false, 00:10:40.752 "nvme_io_md": false, 00:10:40.752 "write_zeroes": true, 00:10:40.752 "zcopy": true, 00:10:40.752 "get_zone_info": false, 00:10:40.752 "zone_management": false, 00:10:40.752 "zone_append": false, 00:10:40.752 "compare": false, 00:10:40.752 "compare_and_write": false, 00:10:40.752 "abort": true, 00:10:40.752 "seek_hole": false, 00:10:40.752 "seek_data": false, 00:10:40.752 "copy": true, 00:10:40.752 "nvme_iov_md": false 00:10:40.752 }, 00:10:40.752 "memory_domains": [ 00:10:40.752 { 00:10:40.752 "dma_device_id": "system", 00:10:40.752 "dma_device_type": 1 00:10:40.752 }, 00:10:40.752 { 00:10:40.752 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:40.752 "dma_device_type": 2 00:10:40.752 } 00:10:40.752 ], 00:10:40.752 "driver_specific": {} 00:10:40.752 } 00:10:40.752 ] 00:10:40.752 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.752 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:40.752 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:40.752 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:40.752 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:40.752 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:40.752 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:40.752 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:40.752 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:40.752 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:40.752 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:40.752 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:40.752 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:40.752 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:40.752 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:40.752 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.752 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:40.752 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.752 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.752 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:40.752 "name": "Existed_Raid", 00:10:40.752 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:40.752 "strip_size_kb": 0, 00:10:40.752 "state": "configuring", 00:10:40.752 "raid_level": "raid1", 00:10:40.752 "superblock": false, 00:10:40.752 "num_base_bdevs": 4, 00:10:40.752 "num_base_bdevs_discovered": 3, 00:10:40.752 "num_base_bdevs_operational": 4, 00:10:40.752 "base_bdevs_list": [ 00:10:40.752 { 00:10:40.752 "name": "BaseBdev1", 00:10:40.752 "uuid": "b50ed93b-bda1-45de-92c7-9eeb2ca10585", 00:10:40.752 "is_configured": true, 00:10:40.752 "data_offset": 0, 00:10:40.752 "data_size": 65536 00:10:40.752 }, 00:10:40.752 { 00:10:40.752 "name": "BaseBdev2", 00:10:40.752 "uuid": "f79959fd-9826-4204-82e0-64d6d4df0881", 00:10:40.752 "is_configured": true, 00:10:40.752 "data_offset": 0, 00:10:40.752 "data_size": 65536 00:10:40.752 }, 00:10:40.752 { 00:10:40.752 "name": "BaseBdev3", 00:10:40.752 "uuid": "c28b01c2-0a8c-42f5-becb-386c5ccfd2ec", 00:10:40.752 "is_configured": true, 00:10:40.752 "data_offset": 0, 00:10:40.752 "data_size": 65536 00:10:40.752 }, 00:10:40.752 { 00:10:40.752 "name": "BaseBdev4", 00:10:40.752 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:40.752 "is_configured": false, 00:10:40.752 "data_offset": 0, 00:10:40.752 "data_size": 0 00:10:40.752 } 00:10:40.752 ] 00:10:40.752 }' 00:10:40.752 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:40.752 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.321 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:41.321 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.321 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.321 [2024-11-27 15:15:09.164342] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:41.321 [2024-11-27 15:15:09.164394] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:10:41.322 [2024-11-27 15:15:09.164403] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:10:41.322 BaseBdev4 00:10:41.322 [2024-11-27 15:15:09.164678] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:10:41.322 [2024-11-27 15:15:09.164823] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:10:41.322 [2024-11-27 15:15:09.164840] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:10:41.322 [2024-11-27 15:15:09.165058] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.322 [ 00:10:41.322 { 00:10:41.322 "name": "BaseBdev4", 00:10:41.322 "aliases": [ 00:10:41.322 "a2c68778-65b2-4fda-af2d-f57c3554e30c" 00:10:41.322 ], 00:10:41.322 "product_name": "Malloc disk", 00:10:41.322 "block_size": 512, 00:10:41.322 "num_blocks": 65536, 00:10:41.322 "uuid": "a2c68778-65b2-4fda-af2d-f57c3554e30c", 00:10:41.322 "assigned_rate_limits": { 00:10:41.322 "rw_ios_per_sec": 0, 00:10:41.322 "rw_mbytes_per_sec": 0, 00:10:41.322 "r_mbytes_per_sec": 0, 00:10:41.322 "w_mbytes_per_sec": 0 00:10:41.322 }, 00:10:41.322 "claimed": true, 00:10:41.322 "claim_type": "exclusive_write", 00:10:41.322 "zoned": false, 00:10:41.322 "supported_io_types": { 00:10:41.322 "read": true, 00:10:41.322 "write": true, 00:10:41.322 "unmap": true, 00:10:41.322 "flush": true, 00:10:41.322 "reset": true, 00:10:41.322 "nvme_admin": false, 00:10:41.322 "nvme_io": false, 00:10:41.322 "nvme_io_md": false, 00:10:41.322 "write_zeroes": true, 00:10:41.322 "zcopy": true, 00:10:41.322 "get_zone_info": false, 00:10:41.322 "zone_management": false, 00:10:41.322 "zone_append": false, 00:10:41.322 "compare": false, 00:10:41.322 "compare_and_write": false, 00:10:41.322 "abort": true, 00:10:41.322 "seek_hole": false, 00:10:41.322 "seek_data": false, 00:10:41.322 "copy": true, 00:10:41.322 "nvme_iov_md": false 00:10:41.322 }, 00:10:41.322 "memory_domains": [ 00:10:41.322 { 00:10:41.322 "dma_device_id": "system", 00:10:41.322 "dma_device_type": 1 00:10:41.322 }, 00:10:41.322 { 00:10:41.322 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:41.322 "dma_device_type": 2 00:10:41.322 } 00:10:41.322 ], 00:10:41.322 "driver_specific": {} 00:10:41.322 } 00:10:41.322 ] 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:41.322 "name": "Existed_Raid", 00:10:41.322 "uuid": "3c198cd1-3949-4e04-b9c8-f5fff3a9d310", 00:10:41.322 "strip_size_kb": 0, 00:10:41.322 "state": "online", 00:10:41.322 "raid_level": "raid1", 00:10:41.322 "superblock": false, 00:10:41.322 "num_base_bdevs": 4, 00:10:41.322 "num_base_bdevs_discovered": 4, 00:10:41.322 "num_base_bdevs_operational": 4, 00:10:41.322 "base_bdevs_list": [ 00:10:41.322 { 00:10:41.322 "name": "BaseBdev1", 00:10:41.322 "uuid": "b50ed93b-bda1-45de-92c7-9eeb2ca10585", 00:10:41.322 "is_configured": true, 00:10:41.322 "data_offset": 0, 00:10:41.322 "data_size": 65536 00:10:41.322 }, 00:10:41.322 { 00:10:41.322 "name": "BaseBdev2", 00:10:41.322 "uuid": "f79959fd-9826-4204-82e0-64d6d4df0881", 00:10:41.322 "is_configured": true, 00:10:41.322 "data_offset": 0, 00:10:41.322 "data_size": 65536 00:10:41.322 }, 00:10:41.322 { 00:10:41.322 "name": "BaseBdev3", 00:10:41.322 "uuid": "c28b01c2-0a8c-42f5-becb-386c5ccfd2ec", 00:10:41.322 "is_configured": true, 00:10:41.322 "data_offset": 0, 00:10:41.322 "data_size": 65536 00:10:41.322 }, 00:10:41.322 { 00:10:41.322 "name": "BaseBdev4", 00:10:41.322 "uuid": "a2c68778-65b2-4fda-af2d-f57c3554e30c", 00:10:41.322 "is_configured": true, 00:10:41.322 "data_offset": 0, 00:10:41.322 "data_size": 65536 00:10:41.322 } 00:10:41.322 ] 00:10:41.322 }' 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:41.322 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.582 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:41.582 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:41.582 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:41.582 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:41.582 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:41.582 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:41.582 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:41.582 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:41.582 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.582 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.582 [2024-11-27 15:15:09.632044] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:41.582 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.582 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:41.582 "name": "Existed_Raid", 00:10:41.582 "aliases": [ 00:10:41.582 "3c198cd1-3949-4e04-b9c8-f5fff3a9d310" 00:10:41.582 ], 00:10:41.582 "product_name": "Raid Volume", 00:10:41.582 "block_size": 512, 00:10:41.582 "num_blocks": 65536, 00:10:41.582 "uuid": "3c198cd1-3949-4e04-b9c8-f5fff3a9d310", 00:10:41.582 "assigned_rate_limits": { 00:10:41.582 "rw_ios_per_sec": 0, 00:10:41.582 "rw_mbytes_per_sec": 0, 00:10:41.582 "r_mbytes_per_sec": 0, 00:10:41.582 "w_mbytes_per_sec": 0 00:10:41.582 }, 00:10:41.582 "claimed": false, 00:10:41.582 "zoned": false, 00:10:41.582 "supported_io_types": { 00:10:41.582 "read": true, 00:10:41.582 "write": true, 00:10:41.582 "unmap": false, 00:10:41.582 "flush": false, 00:10:41.582 "reset": true, 00:10:41.582 "nvme_admin": false, 00:10:41.582 "nvme_io": false, 00:10:41.582 "nvme_io_md": false, 00:10:41.582 "write_zeroes": true, 00:10:41.582 "zcopy": false, 00:10:41.582 "get_zone_info": false, 00:10:41.582 "zone_management": false, 00:10:41.582 "zone_append": false, 00:10:41.582 "compare": false, 00:10:41.582 "compare_and_write": false, 00:10:41.582 "abort": false, 00:10:41.582 "seek_hole": false, 00:10:41.582 "seek_data": false, 00:10:41.582 "copy": false, 00:10:41.582 "nvme_iov_md": false 00:10:41.582 }, 00:10:41.582 "memory_domains": [ 00:10:41.582 { 00:10:41.582 "dma_device_id": "system", 00:10:41.582 "dma_device_type": 1 00:10:41.582 }, 00:10:41.582 { 00:10:41.582 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:41.582 "dma_device_type": 2 00:10:41.582 }, 00:10:41.582 { 00:10:41.582 "dma_device_id": "system", 00:10:41.582 "dma_device_type": 1 00:10:41.582 }, 00:10:41.582 { 00:10:41.582 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:41.582 "dma_device_type": 2 00:10:41.582 }, 00:10:41.582 { 00:10:41.582 "dma_device_id": "system", 00:10:41.582 "dma_device_type": 1 00:10:41.582 }, 00:10:41.582 { 00:10:41.582 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:41.582 "dma_device_type": 2 00:10:41.582 }, 00:10:41.582 { 00:10:41.582 "dma_device_id": "system", 00:10:41.582 "dma_device_type": 1 00:10:41.582 }, 00:10:41.582 { 00:10:41.582 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:41.582 "dma_device_type": 2 00:10:41.582 } 00:10:41.582 ], 00:10:41.582 "driver_specific": { 00:10:41.582 "raid": { 00:10:41.582 "uuid": "3c198cd1-3949-4e04-b9c8-f5fff3a9d310", 00:10:41.583 "strip_size_kb": 0, 00:10:41.583 "state": "online", 00:10:41.583 "raid_level": "raid1", 00:10:41.583 "superblock": false, 00:10:41.583 "num_base_bdevs": 4, 00:10:41.583 "num_base_bdevs_discovered": 4, 00:10:41.583 "num_base_bdevs_operational": 4, 00:10:41.583 "base_bdevs_list": [ 00:10:41.583 { 00:10:41.583 "name": "BaseBdev1", 00:10:41.583 "uuid": "b50ed93b-bda1-45de-92c7-9eeb2ca10585", 00:10:41.583 "is_configured": true, 00:10:41.583 "data_offset": 0, 00:10:41.583 "data_size": 65536 00:10:41.583 }, 00:10:41.583 { 00:10:41.583 "name": "BaseBdev2", 00:10:41.583 "uuid": "f79959fd-9826-4204-82e0-64d6d4df0881", 00:10:41.583 "is_configured": true, 00:10:41.583 "data_offset": 0, 00:10:41.583 "data_size": 65536 00:10:41.583 }, 00:10:41.583 { 00:10:41.583 "name": "BaseBdev3", 00:10:41.583 "uuid": "c28b01c2-0a8c-42f5-becb-386c5ccfd2ec", 00:10:41.583 "is_configured": true, 00:10:41.583 "data_offset": 0, 00:10:41.583 "data_size": 65536 00:10:41.583 }, 00:10:41.583 { 00:10:41.583 "name": "BaseBdev4", 00:10:41.583 "uuid": "a2c68778-65b2-4fda-af2d-f57c3554e30c", 00:10:41.583 "is_configured": true, 00:10:41.583 "data_offset": 0, 00:10:41.583 "data_size": 65536 00:10:41.583 } 00:10:41.583 ] 00:10:41.583 } 00:10:41.583 } 00:10:41.583 }' 00:10:41.583 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:41.843 BaseBdev2 00:10:41.843 BaseBdev3 00:10:41.843 BaseBdev4' 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.843 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.843 [2024-11-27 15:15:09.935222] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:42.103 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.103 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:42.103 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:10:42.103 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:42.103 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:10:42.103 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:10:42.103 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:10:42.103 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:42.103 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:42.103 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:42.103 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:42.103 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:42.103 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:42.103 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:42.103 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:42.103 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:42.103 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:42.103 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:42.103 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.103 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.103 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.103 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:42.103 "name": "Existed_Raid", 00:10:42.103 "uuid": "3c198cd1-3949-4e04-b9c8-f5fff3a9d310", 00:10:42.103 "strip_size_kb": 0, 00:10:42.103 "state": "online", 00:10:42.103 "raid_level": "raid1", 00:10:42.103 "superblock": false, 00:10:42.103 "num_base_bdevs": 4, 00:10:42.103 "num_base_bdevs_discovered": 3, 00:10:42.103 "num_base_bdevs_operational": 3, 00:10:42.103 "base_bdevs_list": [ 00:10:42.103 { 00:10:42.103 "name": null, 00:10:42.103 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:42.103 "is_configured": false, 00:10:42.103 "data_offset": 0, 00:10:42.103 "data_size": 65536 00:10:42.103 }, 00:10:42.103 { 00:10:42.103 "name": "BaseBdev2", 00:10:42.103 "uuid": "f79959fd-9826-4204-82e0-64d6d4df0881", 00:10:42.103 "is_configured": true, 00:10:42.103 "data_offset": 0, 00:10:42.103 "data_size": 65536 00:10:42.103 }, 00:10:42.103 { 00:10:42.103 "name": "BaseBdev3", 00:10:42.103 "uuid": "c28b01c2-0a8c-42f5-becb-386c5ccfd2ec", 00:10:42.103 "is_configured": true, 00:10:42.103 "data_offset": 0, 00:10:42.103 "data_size": 65536 00:10:42.103 }, 00:10:42.103 { 00:10:42.103 "name": "BaseBdev4", 00:10:42.103 "uuid": "a2c68778-65b2-4fda-af2d-f57c3554e30c", 00:10:42.103 "is_configured": true, 00:10:42.103 "data_offset": 0, 00:10:42.103 "data_size": 65536 00:10:42.103 } 00:10:42.103 ] 00:10:42.103 }' 00:10:42.103 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:42.103 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.363 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:42.363 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:42.363 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:42.363 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:42.363 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.363 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.363 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.624 [2024-11-27 15:15:10.481733] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.624 [2024-11-27 15:15:10.549641] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.624 [2024-11-27 15:15:10.616738] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:10:42.624 [2024-11-27 15:15:10.616845] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:42.624 [2024-11-27 15:15:10.628474] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:42.624 [2024-11-27 15:15:10.628527] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:42.624 [2024-11-27 15:15:10.628540] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.624 BaseBdev2 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.624 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.625 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.625 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:42.625 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.625 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.625 [ 00:10:42.625 { 00:10:42.625 "name": "BaseBdev2", 00:10:42.625 "aliases": [ 00:10:42.625 "eb90caa0-7c81-451d-b8ba-e728ee766607" 00:10:42.625 ], 00:10:42.625 "product_name": "Malloc disk", 00:10:42.625 "block_size": 512, 00:10:42.625 "num_blocks": 65536, 00:10:42.625 "uuid": "eb90caa0-7c81-451d-b8ba-e728ee766607", 00:10:42.625 "assigned_rate_limits": { 00:10:42.625 "rw_ios_per_sec": 0, 00:10:42.625 "rw_mbytes_per_sec": 0, 00:10:42.625 "r_mbytes_per_sec": 0, 00:10:42.625 "w_mbytes_per_sec": 0 00:10:42.625 }, 00:10:42.625 "claimed": false, 00:10:42.625 "zoned": false, 00:10:42.625 "supported_io_types": { 00:10:42.625 "read": true, 00:10:42.625 "write": true, 00:10:42.625 "unmap": true, 00:10:42.625 "flush": true, 00:10:42.625 "reset": true, 00:10:42.625 "nvme_admin": false, 00:10:42.625 "nvme_io": false, 00:10:42.625 "nvme_io_md": false, 00:10:42.625 "write_zeroes": true, 00:10:42.625 "zcopy": true, 00:10:42.625 "get_zone_info": false, 00:10:42.625 "zone_management": false, 00:10:42.625 "zone_append": false, 00:10:42.625 "compare": false, 00:10:42.625 "compare_and_write": false, 00:10:42.886 "abort": true, 00:10:42.886 "seek_hole": false, 00:10:42.886 "seek_data": false, 00:10:42.886 "copy": true, 00:10:42.886 "nvme_iov_md": false 00:10:42.886 }, 00:10:42.886 "memory_domains": [ 00:10:42.886 { 00:10:42.886 "dma_device_id": "system", 00:10:42.886 "dma_device_type": 1 00:10:42.886 }, 00:10:42.886 { 00:10:42.886 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:42.886 "dma_device_type": 2 00:10:42.886 } 00:10:42.886 ], 00:10:42.886 "driver_specific": {} 00:10:42.886 } 00:10:42.886 ] 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.886 BaseBdev3 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.886 [ 00:10:42.886 { 00:10:42.886 "name": "BaseBdev3", 00:10:42.886 "aliases": [ 00:10:42.886 "23a4f1d2-8639-4308-b716-53494df33189" 00:10:42.886 ], 00:10:42.886 "product_name": "Malloc disk", 00:10:42.886 "block_size": 512, 00:10:42.886 "num_blocks": 65536, 00:10:42.886 "uuid": "23a4f1d2-8639-4308-b716-53494df33189", 00:10:42.886 "assigned_rate_limits": { 00:10:42.886 "rw_ios_per_sec": 0, 00:10:42.886 "rw_mbytes_per_sec": 0, 00:10:42.886 "r_mbytes_per_sec": 0, 00:10:42.886 "w_mbytes_per_sec": 0 00:10:42.886 }, 00:10:42.886 "claimed": false, 00:10:42.886 "zoned": false, 00:10:42.886 "supported_io_types": { 00:10:42.886 "read": true, 00:10:42.886 "write": true, 00:10:42.886 "unmap": true, 00:10:42.886 "flush": true, 00:10:42.886 "reset": true, 00:10:42.886 "nvme_admin": false, 00:10:42.886 "nvme_io": false, 00:10:42.886 "nvme_io_md": false, 00:10:42.886 "write_zeroes": true, 00:10:42.886 "zcopy": true, 00:10:42.886 "get_zone_info": false, 00:10:42.886 "zone_management": false, 00:10:42.886 "zone_append": false, 00:10:42.886 "compare": false, 00:10:42.886 "compare_and_write": false, 00:10:42.886 "abort": true, 00:10:42.886 "seek_hole": false, 00:10:42.886 "seek_data": false, 00:10:42.886 "copy": true, 00:10:42.886 "nvme_iov_md": false 00:10:42.886 }, 00:10:42.886 "memory_domains": [ 00:10:42.886 { 00:10:42.886 "dma_device_id": "system", 00:10:42.886 "dma_device_type": 1 00:10:42.886 }, 00:10:42.886 { 00:10:42.886 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:42.886 "dma_device_type": 2 00:10:42.886 } 00:10:42.886 ], 00:10:42.886 "driver_specific": {} 00:10:42.886 } 00:10:42.886 ] 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.886 BaseBdev4 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.886 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.886 [ 00:10:42.886 { 00:10:42.886 "name": "BaseBdev4", 00:10:42.886 "aliases": [ 00:10:42.886 "cc7dcb28-4fd5-42aa-b560-4a271d9ea824" 00:10:42.886 ], 00:10:42.886 "product_name": "Malloc disk", 00:10:42.886 "block_size": 512, 00:10:42.886 "num_blocks": 65536, 00:10:42.886 "uuid": "cc7dcb28-4fd5-42aa-b560-4a271d9ea824", 00:10:42.886 "assigned_rate_limits": { 00:10:42.886 "rw_ios_per_sec": 0, 00:10:42.886 "rw_mbytes_per_sec": 0, 00:10:42.886 "r_mbytes_per_sec": 0, 00:10:42.886 "w_mbytes_per_sec": 0 00:10:42.886 }, 00:10:42.886 "claimed": false, 00:10:42.886 "zoned": false, 00:10:42.886 "supported_io_types": { 00:10:42.886 "read": true, 00:10:42.886 "write": true, 00:10:42.887 "unmap": true, 00:10:42.887 "flush": true, 00:10:42.887 "reset": true, 00:10:42.887 "nvme_admin": false, 00:10:42.887 "nvme_io": false, 00:10:42.887 "nvme_io_md": false, 00:10:42.887 "write_zeroes": true, 00:10:42.887 "zcopy": true, 00:10:42.887 "get_zone_info": false, 00:10:42.887 "zone_management": false, 00:10:42.887 "zone_append": false, 00:10:42.887 "compare": false, 00:10:42.887 "compare_and_write": false, 00:10:42.887 "abort": true, 00:10:42.887 "seek_hole": false, 00:10:42.887 "seek_data": false, 00:10:42.887 "copy": true, 00:10:42.887 "nvme_iov_md": false 00:10:42.887 }, 00:10:42.887 "memory_domains": [ 00:10:42.887 { 00:10:42.887 "dma_device_id": "system", 00:10:42.887 "dma_device_type": 1 00:10:42.887 }, 00:10:42.887 { 00:10:42.887 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:42.887 "dma_device_type": 2 00:10:42.887 } 00:10:42.887 ], 00:10:42.887 "driver_specific": {} 00:10:42.887 } 00:10:42.887 ] 00:10:42.887 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.887 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:42.887 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:42.887 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:42.887 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:42.887 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.887 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.887 [2024-11-27 15:15:10.845429] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:42.887 [2024-11-27 15:15:10.845487] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:42.887 [2024-11-27 15:15:10.845506] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:42.887 [2024-11-27 15:15:10.847290] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:42.887 [2024-11-27 15:15:10.847333] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:42.887 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.887 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:42.887 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:42.887 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:42.887 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:42.887 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:42.887 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:42.887 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:42.887 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:42.887 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:42.887 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:42.887 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:42.887 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:42.887 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.887 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.887 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.887 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:42.887 "name": "Existed_Raid", 00:10:42.887 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:42.887 "strip_size_kb": 0, 00:10:42.887 "state": "configuring", 00:10:42.887 "raid_level": "raid1", 00:10:42.887 "superblock": false, 00:10:42.887 "num_base_bdevs": 4, 00:10:42.887 "num_base_bdevs_discovered": 3, 00:10:42.887 "num_base_bdevs_operational": 4, 00:10:42.887 "base_bdevs_list": [ 00:10:42.887 { 00:10:42.887 "name": "BaseBdev1", 00:10:42.887 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:42.887 "is_configured": false, 00:10:42.887 "data_offset": 0, 00:10:42.887 "data_size": 0 00:10:42.887 }, 00:10:42.887 { 00:10:42.887 "name": "BaseBdev2", 00:10:42.887 "uuid": "eb90caa0-7c81-451d-b8ba-e728ee766607", 00:10:42.887 "is_configured": true, 00:10:42.887 "data_offset": 0, 00:10:42.887 "data_size": 65536 00:10:42.887 }, 00:10:42.887 { 00:10:42.887 "name": "BaseBdev3", 00:10:42.887 "uuid": "23a4f1d2-8639-4308-b716-53494df33189", 00:10:42.887 "is_configured": true, 00:10:42.887 "data_offset": 0, 00:10:42.887 "data_size": 65536 00:10:42.887 }, 00:10:42.887 { 00:10:42.887 "name": "BaseBdev4", 00:10:42.887 "uuid": "cc7dcb28-4fd5-42aa-b560-4a271d9ea824", 00:10:42.887 "is_configured": true, 00:10:42.887 "data_offset": 0, 00:10:42.887 "data_size": 65536 00:10:42.887 } 00:10:42.887 ] 00:10:42.887 }' 00:10:42.887 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:42.887 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.457 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:43.457 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:43.457 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.457 [2024-11-27 15:15:11.288727] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:43.457 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:43.457 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:43.457 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:43.457 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:43.457 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:43.457 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:43.457 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:43.457 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:43.457 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:43.457 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:43.457 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:43.457 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:43.457 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:43.457 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:43.457 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.457 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:43.457 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:43.457 "name": "Existed_Raid", 00:10:43.457 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:43.457 "strip_size_kb": 0, 00:10:43.457 "state": "configuring", 00:10:43.457 "raid_level": "raid1", 00:10:43.457 "superblock": false, 00:10:43.457 "num_base_bdevs": 4, 00:10:43.457 "num_base_bdevs_discovered": 2, 00:10:43.457 "num_base_bdevs_operational": 4, 00:10:43.457 "base_bdevs_list": [ 00:10:43.457 { 00:10:43.457 "name": "BaseBdev1", 00:10:43.457 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:43.457 "is_configured": false, 00:10:43.457 "data_offset": 0, 00:10:43.457 "data_size": 0 00:10:43.457 }, 00:10:43.457 { 00:10:43.457 "name": null, 00:10:43.457 "uuid": "eb90caa0-7c81-451d-b8ba-e728ee766607", 00:10:43.457 "is_configured": false, 00:10:43.457 "data_offset": 0, 00:10:43.457 "data_size": 65536 00:10:43.457 }, 00:10:43.457 { 00:10:43.457 "name": "BaseBdev3", 00:10:43.457 "uuid": "23a4f1d2-8639-4308-b716-53494df33189", 00:10:43.457 "is_configured": true, 00:10:43.457 "data_offset": 0, 00:10:43.457 "data_size": 65536 00:10:43.457 }, 00:10:43.457 { 00:10:43.457 "name": "BaseBdev4", 00:10:43.457 "uuid": "cc7dcb28-4fd5-42aa-b560-4a271d9ea824", 00:10:43.457 "is_configured": true, 00:10:43.457 "data_offset": 0, 00:10:43.457 "data_size": 65536 00:10:43.457 } 00:10:43.457 ] 00:10:43.457 }' 00:10:43.457 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:43.457 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.716 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:43.716 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:43.716 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.716 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:43.716 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:43.716 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:43.716 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:43.716 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:43.716 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.716 BaseBdev1 00:10:43.716 [2024-11-27 15:15:11.786778] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:43.716 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:43.716 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:43.716 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:43.716 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:43.716 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:43.716 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:43.716 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:43.716 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:43.716 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:43.716 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.716 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:43.716 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:43.716 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:43.716 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.716 [ 00:10:43.716 { 00:10:43.716 "name": "BaseBdev1", 00:10:43.716 "aliases": [ 00:10:43.716 "af18b0ae-67ad-447d-9f20-63735c2e8b4a" 00:10:43.716 ], 00:10:43.716 "product_name": "Malloc disk", 00:10:43.716 "block_size": 512, 00:10:43.716 "num_blocks": 65536, 00:10:43.716 "uuid": "af18b0ae-67ad-447d-9f20-63735c2e8b4a", 00:10:43.716 "assigned_rate_limits": { 00:10:43.716 "rw_ios_per_sec": 0, 00:10:43.716 "rw_mbytes_per_sec": 0, 00:10:43.716 "r_mbytes_per_sec": 0, 00:10:43.716 "w_mbytes_per_sec": 0 00:10:43.716 }, 00:10:43.716 "claimed": true, 00:10:43.716 "claim_type": "exclusive_write", 00:10:43.716 "zoned": false, 00:10:43.716 "supported_io_types": { 00:10:43.716 "read": true, 00:10:43.716 "write": true, 00:10:43.716 "unmap": true, 00:10:43.716 "flush": true, 00:10:43.716 "reset": true, 00:10:43.716 "nvme_admin": false, 00:10:43.716 "nvme_io": false, 00:10:43.716 "nvme_io_md": false, 00:10:43.716 "write_zeroes": true, 00:10:43.716 "zcopy": true, 00:10:43.716 "get_zone_info": false, 00:10:43.716 "zone_management": false, 00:10:43.716 "zone_append": false, 00:10:43.716 "compare": false, 00:10:43.716 "compare_and_write": false, 00:10:43.716 "abort": true, 00:10:43.716 "seek_hole": false, 00:10:43.716 "seek_data": false, 00:10:43.716 "copy": true, 00:10:43.716 "nvme_iov_md": false 00:10:43.976 }, 00:10:43.976 "memory_domains": [ 00:10:43.976 { 00:10:43.976 "dma_device_id": "system", 00:10:43.976 "dma_device_type": 1 00:10:43.976 }, 00:10:43.976 { 00:10:43.976 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:43.976 "dma_device_type": 2 00:10:43.976 } 00:10:43.976 ], 00:10:43.976 "driver_specific": {} 00:10:43.976 } 00:10:43.976 ] 00:10:43.976 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:43.976 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:43.976 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:43.976 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:43.976 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:43.976 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:43.976 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:43.976 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:43.976 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:43.976 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:43.976 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:43.976 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:43.976 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:43.976 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:43.976 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:43.976 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.976 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:43.976 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:43.976 "name": "Existed_Raid", 00:10:43.976 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:43.976 "strip_size_kb": 0, 00:10:43.976 "state": "configuring", 00:10:43.976 "raid_level": "raid1", 00:10:43.976 "superblock": false, 00:10:43.976 "num_base_bdevs": 4, 00:10:43.976 "num_base_bdevs_discovered": 3, 00:10:43.976 "num_base_bdevs_operational": 4, 00:10:43.976 "base_bdevs_list": [ 00:10:43.976 { 00:10:43.976 "name": "BaseBdev1", 00:10:43.976 "uuid": "af18b0ae-67ad-447d-9f20-63735c2e8b4a", 00:10:43.976 "is_configured": true, 00:10:43.976 "data_offset": 0, 00:10:43.976 "data_size": 65536 00:10:43.976 }, 00:10:43.976 { 00:10:43.976 "name": null, 00:10:43.976 "uuid": "eb90caa0-7c81-451d-b8ba-e728ee766607", 00:10:43.976 "is_configured": false, 00:10:43.976 "data_offset": 0, 00:10:43.976 "data_size": 65536 00:10:43.976 }, 00:10:43.976 { 00:10:43.976 "name": "BaseBdev3", 00:10:43.976 "uuid": "23a4f1d2-8639-4308-b716-53494df33189", 00:10:43.976 "is_configured": true, 00:10:43.976 "data_offset": 0, 00:10:43.976 "data_size": 65536 00:10:43.976 }, 00:10:43.976 { 00:10:43.976 "name": "BaseBdev4", 00:10:43.976 "uuid": "cc7dcb28-4fd5-42aa-b560-4a271d9ea824", 00:10:43.976 "is_configured": true, 00:10:43.976 "data_offset": 0, 00:10:43.976 "data_size": 65536 00:10:43.976 } 00:10:43.976 ] 00:10:43.976 }' 00:10:43.976 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:43.976 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.236 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:44.236 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.236 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:44.236 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.236 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.236 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:44.236 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:44.236 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.236 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.236 [2024-11-27 15:15:12.321953] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:44.236 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.236 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:44.236 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:44.236 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:44.236 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:44.236 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:44.236 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:44.236 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:44.236 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:44.236 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:44.236 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:44.236 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:44.236 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:44.236 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.236 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.496 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.496 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:44.496 "name": "Existed_Raid", 00:10:44.496 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:44.496 "strip_size_kb": 0, 00:10:44.496 "state": "configuring", 00:10:44.496 "raid_level": "raid1", 00:10:44.496 "superblock": false, 00:10:44.496 "num_base_bdevs": 4, 00:10:44.496 "num_base_bdevs_discovered": 2, 00:10:44.496 "num_base_bdevs_operational": 4, 00:10:44.496 "base_bdevs_list": [ 00:10:44.496 { 00:10:44.496 "name": "BaseBdev1", 00:10:44.496 "uuid": "af18b0ae-67ad-447d-9f20-63735c2e8b4a", 00:10:44.496 "is_configured": true, 00:10:44.496 "data_offset": 0, 00:10:44.496 "data_size": 65536 00:10:44.496 }, 00:10:44.496 { 00:10:44.496 "name": null, 00:10:44.496 "uuid": "eb90caa0-7c81-451d-b8ba-e728ee766607", 00:10:44.496 "is_configured": false, 00:10:44.496 "data_offset": 0, 00:10:44.496 "data_size": 65536 00:10:44.496 }, 00:10:44.496 { 00:10:44.496 "name": null, 00:10:44.496 "uuid": "23a4f1d2-8639-4308-b716-53494df33189", 00:10:44.496 "is_configured": false, 00:10:44.496 "data_offset": 0, 00:10:44.496 "data_size": 65536 00:10:44.496 }, 00:10:44.496 { 00:10:44.496 "name": "BaseBdev4", 00:10:44.496 "uuid": "cc7dcb28-4fd5-42aa-b560-4a271d9ea824", 00:10:44.496 "is_configured": true, 00:10:44.496 "data_offset": 0, 00:10:44.496 "data_size": 65536 00:10:44.496 } 00:10:44.496 ] 00:10:44.496 }' 00:10:44.496 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:44.496 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.756 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:44.756 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:44.756 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.756 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.756 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.756 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:44.756 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:44.756 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.756 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.756 [2024-11-27 15:15:12.809176] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:44.756 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.756 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:44.756 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:44.756 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:44.756 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:44.756 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:44.756 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:44.756 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:44.756 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:44.756 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:44.756 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:44.756 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:44.756 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:44.756 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.756 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.756 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:45.015 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:45.015 "name": "Existed_Raid", 00:10:45.015 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:45.015 "strip_size_kb": 0, 00:10:45.015 "state": "configuring", 00:10:45.015 "raid_level": "raid1", 00:10:45.015 "superblock": false, 00:10:45.015 "num_base_bdevs": 4, 00:10:45.015 "num_base_bdevs_discovered": 3, 00:10:45.015 "num_base_bdevs_operational": 4, 00:10:45.015 "base_bdevs_list": [ 00:10:45.015 { 00:10:45.015 "name": "BaseBdev1", 00:10:45.015 "uuid": "af18b0ae-67ad-447d-9f20-63735c2e8b4a", 00:10:45.015 "is_configured": true, 00:10:45.015 "data_offset": 0, 00:10:45.015 "data_size": 65536 00:10:45.015 }, 00:10:45.015 { 00:10:45.015 "name": null, 00:10:45.015 "uuid": "eb90caa0-7c81-451d-b8ba-e728ee766607", 00:10:45.015 "is_configured": false, 00:10:45.015 "data_offset": 0, 00:10:45.015 "data_size": 65536 00:10:45.015 }, 00:10:45.015 { 00:10:45.015 "name": "BaseBdev3", 00:10:45.015 "uuid": "23a4f1d2-8639-4308-b716-53494df33189", 00:10:45.015 "is_configured": true, 00:10:45.015 "data_offset": 0, 00:10:45.015 "data_size": 65536 00:10:45.015 }, 00:10:45.015 { 00:10:45.015 "name": "BaseBdev4", 00:10:45.015 "uuid": "cc7dcb28-4fd5-42aa-b560-4a271d9ea824", 00:10:45.015 "is_configured": true, 00:10:45.015 "data_offset": 0, 00:10:45.015 "data_size": 65536 00:10:45.015 } 00:10:45.015 ] 00:10:45.015 }' 00:10:45.015 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:45.015 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.294 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:45.294 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:45.294 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.294 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:45.294 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:45.294 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:45.294 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:45.294 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:45.294 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.294 [2024-11-27 15:15:13.296303] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:45.294 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:45.294 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:45.294 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:45.294 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:45.294 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:45.294 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:45.294 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:45.294 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:45.294 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:45.294 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:45.294 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:45.294 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:45.294 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:45.294 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.294 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:45.294 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:45.294 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:45.294 "name": "Existed_Raid", 00:10:45.294 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:45.294 "strip_size_kb": 0, 00:10:45.294 "state": "configuring", 00:10:45.294 "raid_level": "raid1", 00:10:45.294 "superblock": false, 00:10:45.294 "num_base_bdevs": 4, 00:10:45.294 "num_base_bdevs_discovered": 2, 00:10:45.294 "num_base_bdevs_operational": 4, 00:10:45.294 "base_bdevs_list": [ 00:10:45.294 { 00:10:45.294 "name": null, 00:10:45.294 "uuid": "af18b0ae-67ad-447d-9f20-63735c2e8b4a", 00:10:45.294 "is_configured": false, 00:10:45.294 "data_offset": 0, 00:10:45.294 "data_size": 65536 00:10:45.294 }, 00:10:45.294 { 00:10:45.294 "name": null, 00:10:45.294 "uuid": "eb90caa0-7c81-451d-b8ba-e728ee766607", 00:10:45.294 "is_configured": false, 00:10:45.294 "data_offset": 0, 00:10:45.294 "data_size": 65536 00:10:45.294 }, 00:10:45.294 { 00:10:45.294 "name": "BaseBdev3", 00:10:45.294 "uuid": "23a4f1d2-8639-4308-b716-53494df33189", 00:10:45.294 "is_configured": true, 00:10:45.294 "data_offset": 0, 00:10:45.294 "data_size": 65536 00:10:45.294 }, 00:10:45.294 { 00:10:45.294 "name": "BaseBdev4", 00:10:45.294 "uuid": "cc7dcb28-4fd5-42aa-b560-4a271d9ea824", 00:10:45.294 "is_configured": true, 00:10:45.294 "data_offset": 0, 00:10:45.294 "data_size": 65536 00:10:45.294 } 00:10:45.294 ] 00:10:45.294 }' 00:10:45.294 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:45.294 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.881 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:45.881 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:45.881 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.881 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:45.881 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:45.881 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:45.881 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:45.882 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:45.882 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.882 [2024-11-27 15:15:13.786418] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:45.882 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:45.882 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:45.882 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:45.882 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:45.882 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:45.882 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:45.882 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:45.882 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:45.882 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:45.882 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:45.882 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:45.882 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:45.882 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:45.882 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.882 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:45.882 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:45.882 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:45.882 "name": "Existed_Raid", 00:10:45.882 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:45.882 "strip_size_kb": 0, 00:10:45.882 "state": "configuring", 00:10:45.882 "raid_level": "raid1", 00:10:45.882 "superblock": false, 00:10:45.882 "num_base_bdevs": 4, 00:10:45.882 "num_base_bdevs_discovered": 3, 00:10:45.882 "num_base_bdevs_operational": 4, 00:10:45.882 "base_bdevs_list": [ 00:10:45.882 { 00:10:45.882 "name": null, 00:10:45.882 "uuid": "af18b0ae-67ad-447d-9f20-63735c2e8b4a", 00:10:45.882 "is_configured": false, 00:10:45.882 "data_offset": 0, 00:10:45.882 "data_size": 65536 00:10:45.882 }, 00:10:45.882 { 00:10:45.882 "name": "BaseBdev2", 00:10:45.882 "uuid": "eb90caa0-7c81-451d-b8ba-e728ee766607", 00:10:45.882 "is_configured": true, 00:10:45.882 "data_offset": 0, 00:10:45.882 "data_size": 65536 00:10:45.882 }, 00:10:45.882 { 00:10:45.882 "name": "BaseBdev3", 00:10:45.882 "uuid": "23a4f1d2-8639-4308-b716-53494df33189", 00:10:45.882 "is_configured": true, 00:10:45.882 "data_offset": 0, 00:10:45.882 "data_size": 65536 00:10:45.882 }, 00:10:45.882 { 00:10:45.882 "name": "BaseBdev4", 00:10:45.882 "uuid": "cc7dcb28-4fd5-42aa-b560-4a271d9ea824", 00:10:45.882 "is_configured": true, 00:10:45.882 "data_offset": 0, 00:10:45.882 "data_size": 65536 00:10:45.882 } 00:10:45.882 ] 00:10:45.882 }' 00:10:45.882 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:45.882 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.451 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:46.451 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.451 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.451 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:46.451 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.451 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:46.451 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:46.451 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:46.451 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.451 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.451 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.451 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u af18b0ae-67ad-447d-9f20-63735c2e8b4a 00:10:46.451 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.451 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.451 [2024-11-27 15:15:14.368572] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:46.451 [2024-11-27 15:15:14.368715] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:10:46.451 [2024-11-27 15:15:14.368751] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:10:46.451 [2024-11-27 15:15:14.369066] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:10:46.451 [2024-11-27 15:15:14.369235] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:10:46.451 [2024-11-27 15:15:14.369276] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006d00 00:10:46.451 [2024-11-27 15:15:14.369522] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:46.451 NewBaseBdev 00:10:46.451 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.451 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:46.451 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:10:46.451 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:46.451 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:46.451 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:46.451 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:46.451 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:46.451 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.451 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.451 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.451 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:46.451 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.451 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.451 [ 00:10:46.451 { 00:10:46.451 "name": "NewBaseBdev", 00:10:46.452 "aliases": [ 00:10:46.452 "af18b0ae-67ad-447d-9f20-63735c2e8b4a" 00:10:46.452 ], 00:10:46.452 "product_name": "Malloc disk", 00:10:46.452 "block_size": 512, 00:10:46.452 "num_blocks": 65536, 00:10:46.452 "uuid": "af18b0ae-67ad-447d-9f20-63735c2e8b4a", 00:10:46.452 "assigned_rate_limits": { 00:10:46.452 "rw_ios_per_sec": 0, 00:10:46.452 "rw_mbytes_per_sec": 0, 00:10:46.452 "r_mbytes_per_sec": 0, 00:10:46.452 "w_mbytes_per_sec": 0 00:10:46.452 }, 00:10:46.452 "claimed": true, 00:10:46.452 "claim_type": "exclusive_write", 00:10:46.452 "zoned": false, 00:10:46.452 "supported_io_types": { 00:10:46.452 "read": true, 00:10:46.452 "write": true, 00:10:46.452 "unmap": true, 00:10:46.452 "flush": true, 00:10:46.452 "reset": true, 00:10:46.452 "nvme_admin": false, 00:10:46.452 "nvme_io": false, 00:10:46.452 "nvme_io_md": false, 00:10:46.452 "write_zeroes": true, 00:10:46.452 "zcopy": true, 00:10:46.452 "get_zone_info": false, 00:10:46.452 "zone_management": false, 00:10:46.452 "zone_append": false, 00:10:46.452 "compare": false, 00:10:46.452 "compare_and_write": false, 00:10:46.452 "abort": true, 00:10:46.452 "seek_hole": false, 00:10:46.452 "seek_data": false, 00:10:46.452 "copy": true, 00:10:46.452 "nvme_iov_md": false 00:10:46.452 }, 00:10:46.452 "memory_domains": [ 00:10:46.452 { 00:10:46.452 "dma_device_id": "system", 00:10:46.452 "dma_device_type": 1 00:10:46.452 }, 00:10:46.452 { 00:10:46.452 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:46.452 "dma_device_type": 2 00:10:46.452 } 00:10:46.452 ], 00:10:46.452 "driver_specific": {} 00:10:46.452 } 00:10:46.452 ] 00:10:46.452 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.452 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:46.452 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:10:46.452 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:46.452 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:46.452 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:46.452 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:46.452 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:46.452 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:46.452 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:46.452 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:46.452 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:46.452 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:46.452 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:46.452 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.452 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.452 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.452 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:46.452 "name": "Existed_Raid", 00:10:46.452 "uuid": "7ce966a9-839e-4fb1-929c-03699fc70d9d", 00:10:46.452 "strip_size_kb": 0, 00:10:46.452 "state": "online", 00:10:46.452 "raid_level": "raid1", 00:10:46.452 "superblock": false, 00:10:46.452 "num_base_bdevs": 4, 00:10:46.452 "num_base_bdevs_discovered": 4, 00:10:46.452 "num_base_bdevs_operational": 4, 00:10:46.452 "base_bdevs_list": [ 00:10:46.452 { 00:10:46.452 "name": "NewBaseBdev", 00:10:46.452 "uuid": "af18b0ae-67ad-447d-9f20-63735c2e8b4a", 00:10:46.452 "is_configured": true, 00:10:46.452 "data_offset": 0, 00:10:46.452 "data_size": 65536 00:10:46.452 }, 00:10:46.452 { 00:10:46.452 "name": "BaseBdev2", 00:10:46.452 "uuid": "eb90caa0-7c81-451d-b8ba-e728ee766607", 00:10:46.452 "is_configured": true, 00:10:46.452 "data_offset": 0, 00:10:46.452 "data_size": 65536 00:10:46.452 }, 00:10:46.452 { 00:10:46.452 "name": "BaseBdev3", 00:10:46.452 "uuid": "23a4f1d2-8639-4308-b716-53494df33189", 00:10:46.452 "is_configured": true, 00:10:46.452 "data_offset": 0, 00:10:46.452 "data_size": 65536 00:10:46.452 }, 00:10:46.452 { 00:10:46.452 "name": "BaseBdev4", 00:10:46.452 "uuid": "cc7dcb28-4fd5-42aa-b560-4a271d9ea824", 00:10:46.452 "is_configured": true, 00:10:46.452 "data_offset": 0, 00:10:46.452 "data_size": 65536 00:10:46.452 } 00:10:46.452 ] 00:10:46.452 }' 00:10:46.452 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:46.452 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.711 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:46.711 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:46.711 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:46.711 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:46.711 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:46.711 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:46.711 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:46.711 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.711 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.711 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:46.711 [2024-11-27 15:15:14.808210] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:46.970 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.970 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:46.970 "name": "Existed_Raid", 00:10:46.970 "aliases": [ 00:10:46.970 "7ce966a9-839e-4fb1-929c-03699fc70d9d" 00:10:46.970 ], 00:10:46.970 "product_name": "Raid Volume", 00:10:46.970 "block_size": 512, 00:10:46.970 "num_blocks": 65536, 00:10:46.970 "uuid": "7ce966a9-839e-4fb1-929c-03699fc70d9d", 00:10:46.970 "assigned_rate_limits": { 00:10:46.970 "rw_ios_per_sec": 0, 00:10:46.970 "rw_mbytes_per_sec": 0, 00:10:46.970 "r_mbytes_per_sec": 0, 00:10:46.970 "w_mbytes_per_sec": 0 00:10:46.970 }, 00:10:46.970 "claimed": false, 00:10:46.970 "zoned": false, 00:10:46.970 "supported_io_types": { 00:10:46.970 "read": true, 00:10:46.970 "write": true, 00:10:46.970 "unmap": false, 00:10:46.970 "flush": false, 00:10:46.970 "reset": true, 00:10:46.970 "nvme_admin": false, 00:10:46.970 "nvme_io": false, 00:10:46.970 "nvme_io_md": false, 00:10:46.970 "write_zeroes": true, 00:10:46.970 "zcopy": false, 00:10:46.970 "get_zone_info": false, 00:10:46.970 "zone_management": false, 00:10:46.970 "zone_append": false, 00:10:46.970 "compare": false, 00:10:46.970 "compare_and_write": false, 00:10:46.970 "abort": false, 00:10:46.970 "seek_hole": false, 00:10:46.970 "seek_data": false, 00:10:46.970 "copy": false, 00:10:46.970 "nvme_iov_md": false 00:10:46.970 }, 00:10:46.970 "memory_domains": [ 00:10:46.970 { 00:10:46.970 "dma_device_id": "system", 00:10:46.970 "dma_device_type": 1 00:10:46.970 }, 00:10:46.970 { 00:10:46.970 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:46.970 "dma_device_type": 2 00:10:46.970 }, 00:10:46.970 { 00:10:46.970 "dma_device_id": "system", 00:10:46.970 "dma_device_type": 1 00:10:46.970 }, 00:10:46.970 { 00:10:46.970 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:46.970 "dma_device_type": 2 00:10:46.970 }, 00:10:46.970 { 00:10:46.970 "dma_device_id": "system", 00:10:46.970 "dma_device_type": 1 00:10:46.970 }, 00:10:46.970 { 00:10:46.970 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:46.970 "dma_device_type": 2 00:10:46.970 }, 00:10:46.970 { 00:10:46.970 "dma_device_id": "system", 00:10:46.970 "dma_device_type": 1 00:10:46.970 }, 00:10:46.970 { 00:10:46.970 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:46.970 "dma_device_type": 2 00:10:46.970 } 00:10:46.970 ], 00:10:46.970 "driver_specific": { 00:10:46.970 "raid": { 00:10:46.970 "uuid": "7ce966a9-839e-4fb1-929c-03699fc70d9d", 00:10:46.970 "strip_size_kb": 0, 00:10:46.970 "state": "online", 00:10:46.970 "raid_level": "raid1", 00:10:46.970 "superblock": false, 00:10:46.970 "num_base_bdevs": 4, 00:10:46.970 "num_base_bdevs_discovered": 4, 00:10:46.970 "num_base_bdevs_operational": 4, 00:10:46.970 "base_bdevs_list": [ 00:10:46.970 { 00:10:46.970 "name": "NewBaseBdev", 00:10:46.970 "uuid": "af18b0ae-67ad-447d-9f20-63735c2e8b4a", 00:10:46.970 "is_configured": true, 00:10:46.970 "data_offset": 0, 00:10:46.970 "data_size": 65536 00:10:46.970 }, 00:10:46.970 { 00:10:46.970 "name": "BaseBdev2", 00:10:46.970 "uuid": "eb90caa0-7c81-451d-b8ba-e728ee766607", 00:10:46.970 "is_configured": true, 00:10:46.970 "data_offset": 0, 00:10:46.970 "data_size": 65536 00:10:46.970 }, 00:10:46.970 { 00:10:46.970 "name": "BaseBdev3", 00:10:46.970 "uuid": "23a4f1d2-8639-4308-b716-53494df33189", 00:10:46.970 "is_configured": true, 00:10:46.970 "data_offset": 0, 00:10:46.970 "data_size": 65536 00:10:46.970 }, 00:10:46.970 { 00:10:46.970 "name": "BaseBdev4", 00:10:46.970 "uuid": "cc7dcb28-4fd5-42aa-b560-4a271d9ea824", 00:10:46.970 "is_configured": true, 00:10:46.970 "data_offset": 0, 00:10:46.970 "data_size": 65536 00:10:46.970 } 00:10:46.970 ] 00:10:46.970 } 00:10:46.970 } 00:10:46.970 }' 00:10:46.970 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:46.970 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:46.970 BaseBdev2 00:10:46.970 BaseBdev3 00:10:46.970 BaseBdev4' 00:10:46.970 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:46.970 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:46.970 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:46.970 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:46.970 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:46.970 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.970 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.970 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.970 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:46.970 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:46.970 15:15:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:46.970 15:15:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:46.970 15:15:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:46.970 15:15:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.970 15:15:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.970 15:15:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.970 15:15:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:46.970 15:15:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:46.970 15:15:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:46.970 15:15:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:46.970 15:15:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:46.970 15:15:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.970 15:15:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:47.229 15:15:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:47.229 15:15:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:47.229 15:15:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:47.229 15:15:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:47.229 15:15:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:47.229 15:15:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:47.229 15:15:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:47.229 15:15:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:47.229 15:15:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:47.229 15:15:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:47.229 15:15:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:47.229 15:15:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:47.229 15:15:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:47.229 15:15:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:47.229 [2024-11-27 15:15:15.155273] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:47.229 [2024-11-27 15:15:15.155302] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:47.229 [2024-11-27 15:15:15.155388] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:47.229 [2024-11-27 15:15:15.155648] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:47.229 [2024-11-27 15:15:15.155664] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name Existed_Raid, state offline 00:10:47.229 15:15:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:47.229 15:15:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 84125 00:10:47.229 15:15:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 84125 ']' 00:10:47.229 15:15:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 84125 00:10:47.229 15:15:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:10:47.229 15:15:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:47.229 15:15:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 84125 00:10:47.229 15:15:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:47.230 15:15:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:47.230 killing process with pid 84125 00:10:47.230 15:15:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 84125' 00:10:47.230 15:15:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 84125 00:10:47.230 [2024-11-27 15:15:15.207259] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:47.230 15:15:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 84125 00:10:47.230 [2024-11-27 15:15:15.249268] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:47.489 15:15:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:10:47.489 00:10:47.489 real 0m9.641s 00:10:47.489 user 0m16.450s 00:10:47.489 sys 0m2.072s 00:10:47.489 15:15:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:47.489 15:15:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:47.489 ************************************ 00:10:47.489 END TEST raid_state_function_test 00:10:47.489 ************************************ 00:10:47.489 15:15:15 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 4 true 00:10:47.489 15:15:15 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:47.489 15:15:15 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:47.489 15:15:15 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:47.489 ************************************ 00:10:47.489 START TEST raid_state_function_test_sb 00:10:47.489 ************************************ 00:10:47.489 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 4 true 00:10:47.489 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:10:47.489 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:10:47.489 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:10:47.489 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:10:47.489 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:10:47.489 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:47.489 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:10:47.489 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:47.489 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:47.489 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:10:47.489 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:47.489 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:47.490 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:10:47.490 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:47.490 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:47.490 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:10:47.490 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:47.490 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:47.490 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:47.490 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:10:47.490 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:10:47.490 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:10:47.490 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:10:47.490 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:10:47.490 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:10:47.490 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:10:47.490 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:10:47.490 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:10:47.490 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=84772 00:10:47.490 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:10:47.490 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 84772' 00:10:47.490 Process raid pid: 84772 00:10:47.490 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 84772 00:10:47.490 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 84772 ']' 00:10:47.490 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:47.490 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:47.490 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:47.490 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:47.490 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:47.490 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.750 [2024-11-27 15:15:15.644546] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:10:47.750 [2024-11-27 15:15:15.644791] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:47.750 [2024-11-27 15:15:15.822031] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:47.750 [2024-11-27 15:15:15.849853] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:48.009 [2024-11-27 15:15:15.892859] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:48.009 [2024-11-27 15:15:15.893012] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:48.576 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:48.576 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:10:48.576 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:48.576 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.576 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.576 [2024-11-27 15:15:16.476270] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:48.576 [2024-11-27 15:15:16.476336] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:48.577 [2024-11-27 15:15:16.476347] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:48.577 [2024-11-27 15:15:16.476356] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:48.577 [2024-11-27 15:15:16.476362] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:48.577 [2024-11-27 15:15:16.476375] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:48.577 [2024-11-27 15:15:16.476383] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:48.577 [2024-11-27 15:15:16.476391] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:48.577 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.577 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:48.577 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:48.577 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:48.577 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:48.577 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:48.577 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:48.577 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:48.577 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:48.577 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:48.577 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:48.577 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:48.577 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:48.577 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.577 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.577 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.577 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:48.577 "name": "Existed_Raid", 00:10:48.577 "uuid": "78e6e302-44fd-4568-b776-a8c769d96120", 00:10:48.577 "strip_size_kb": 0, 00:10:48.577 "state": "configuring", 00:10:48.577 "raid_level": "raid1", 00:10:48.577 "superblock": true, 00:10:48.577 "num_base_bdevs": 4, 00:10:48.577 "num_base_bdevs_discovered": 0, 00:10:48.577 "num_base_bdevs_operational": 4, 00:10:48.577 "base_bdevs_list": [ 00:10:48.577 { 00:10:48.577 "name": "BaseBdev1", 00:10:48.577 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:48.577 "is_configured": false, 00:10:48.577 "data_offset": 0, 00:10:48.577 "data_size": 0 00:10:48.577 }, 00:10:48.577 { 00:10:48.577 "name": "BaseBdev2", 00:10:48.577 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:48.577 "is_configured": false, 00:10:48.577 "data_offset": 0, 00:10:48.577 "data_size": 0 00:10:48.577 }, 00:10:48.577 { 00:10:48.577 "name": "BaseBdev3", 00:10:48.577 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:48.577 "is_configured": false, 00:10:48.577 "data_offset": 0, 00:10:48.577 "data_size": 0 00:10:48.577 }, 00:10:48.577 { 00:10:48.577 "name": "BaseBdev4", 00:10:48.577 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:48.577 "is_configured": false, 00:10:48.577 "data_offset": 0, 00:10:48.577 "data_size": 0 00:10:48.577 } 00:10:48.577 ] 00:10:48.577 }' 00:10:48.577 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:48.577 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.835 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:48.836 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.836 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.836 [2024-11-27 15:15:16.931380] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:48.836 [2024-11-27 15:15:16.931497] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:10:48.836 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.836 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:48.836 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.836 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.836 [2024-11-27 15:15:16.939400] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:48.836 [2024-11-27 15:15:16.939496] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:48.836 [2024-11-27 15:15:16.939531] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:48.836 [2024-11-27 15:15:16.939564] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:48.836 [2024-11-27 15:15:16.939586] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:48.836 [2024-11-27 15:15:16.939609] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:48.836 [2024-11-27 15:15:16.939647] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:48.836 [2024-11-27 15:15:16.939673] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:49.095 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.095 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:49.095 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:49.095 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:49.095 [2024-11-27 15:15:16.956416] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:49.095 BaseBdev1 00:10:49.095 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.095 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:10:49.095 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:49.095 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:49.095 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:49.095 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:49.095 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:49.095 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:49.095 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:49.095 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:49.095 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.095 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:49.095 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:49.095 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:49.095 [ 00:10:49.095 { 00:10:49.095 "name": "BaseBdev1", 00:10:49.095 "aliases": [ 00:10:49.095 "8f008574-465a-4b23-8961-e5b7228b439f" 00:10:49.095 ], 00:10:49.095 "product_name": "Malloc disk", 00:10:49.095 "block_size": 512, 00:10:49.095 "num_blocks": 65536, 00:10:49.095 "uuid": "8f008574-465a-4b23-8961-e5b7228b439f", 00:10:49.095 "assigned_rate_limits": { 00:10:49.095 "rw_ios_per_sec": 0, 00:10:49.095 "rw_mbytes_per_sec": 0, 00:10:49.095 "r_mbytes_per_sec": 0, 00:10:49.095 "w_mbytes_per_sec": 0 00:10:49.095 }, 00:10:49.095 "claimed": true, 00:10:49.095 "claim_type": "exclusive_write", 00:10:49.095 "zoned": false, 00:10:49.095 "supported_io_types": { 00:10:49.095 "read": true, 00:10:49.095 "write": true, 00:10:49.095 "unmap": true, 00:10:49.095 "flush": true, 00:10:49.095 "reset": true, 00:10:49.095 "nvme_admin": false, 00:10:49.095 "nvme_io": false, 00:10:49.095 "nvme_io_md": false, 00:10:49.095 "write_zeroes": true, 00:10:49.095 "zcopy": true, 00:10:49.095 "get_zone_info": false, 00:10:49.095 "zone_management": false, 00:10:49.095 "zone_append": false, 00:10:49.095 "compare": false, 00:10:49.095 "compare_and_write": false, 00:10:49.095 "abort": true, 00:10:49.095 "seek_hole": false, 00:10:49.095 "seek_data": false, 00:10:49.095 "copy": true, 00:10:49.095 "nvme_iov_md": false 00:10:49.095 }, 00:10:49.096 "memory_domains": [ 00:10:49.096 { 00:10:49.096 "dma_device_id": "system", 00:10:49.096 "dma_device_type": 1 00:10:49.096 }, 00:10:49.096 { 00:10:49.096 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:49.096 "dma_device_type": 2 00:10:49.096 } 00:10:49.096 ], 00:10:49.096 "driver_specific": {} 00:10:49.096 } 00:10:49.096 ] 00:10:49.096 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.096 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:49.096 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:49.096 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:49.096 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:49.096 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:49.096 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:49.096 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:49.096 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:49.096 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:49.096 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:49.096 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:49.096 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:49.096 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:49.096 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:49.096 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:49.096 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.096 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:49.096 "name": "Existed_Raid", 00:10:49.096 "uuid": "4d540631-f497-4988-bf1d-dc4425f8745d", 00:10:49.096 "strip_size_kb": 0, 00:10:49.096 "state": "configuring", 00:10:49.096 "raid_level": "raid1", 00:10:49.096 "superblock": true, 00:10:49.096 "num_base_bdevs": 4, 00:10:49.096 "num_base_bdevs_discovered": 1, 00:10:49.096 "num_base_bdevs_operational": 4, 00:10:49.096 "base_bdevs_list": [ 00:10:49.096 { 00:10:49.096 "name": "BaseBdev1", 00:10:49.096 "uuid": "8f008574-465a-4b23-8961-e5b7228b439f", 00:10:49.096 "is_configured": true, 00:10:49.096 "data_offset": 2048, 00:10:49.096 "data_size": 63488 00:10:49.096 }, 00:10:49.096 { 00:10:49.096 "name": "BaseBdev2", 00:10:49.096 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:49.096 "is_configured": false, 00:10:49.096 "data_offset": 0, 00:10:49.096 "data_size": 0 00:10:49.096 }, 00:10:49.096 { 00:10:49.096 "name": "BaseBdev3", 00:10:49.096 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:49.096 "is_configured": false, 00:10:49.096 "data_offset": 0, 00:10:49.096 "data_size": 0 00:10:49.096 }, 00:10:49.096 { 00:10:49.096 "name": "BaseBdev4", 00:10:49.096 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:49.096 "is_configured": false, 00:10:49.096 "data_offset": 0, 00:10:49.096 "data_size": 0 00:10:49.096 } 00:10:49.096 ] 00:10:49.096 }' 00:10:49.096 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:49.096 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:49.356 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:49.356 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:49.356 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:49.356 [2024-11-27 15:15:17.431678] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:49.356 [2024-11-27 15:15:17.431826] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:10:49.356 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.356 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:49.356 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:49.356 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:49.356 [2024-11-27 15:15:17.443680] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:49.356 [2024-11-27 15:15:17.445649] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:49.356 [2024-11-27 15:15:17.445729] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:49.356 [2024-11-27 15:15:17.445763] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:49.356 [2024-11-27 15:15:17.445786] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:49.356 [2024-11-27 15:15:17.445812] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:49.356 [2024-11-27 15:15:17.445834] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:49.356 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.356 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:10:49.356 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:49.357 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:49.357 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:49.357 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:49.357 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:49.357 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:49.357 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:49.357 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:49.357 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:49.357 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:49.357 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:49.357 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:49.357 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:49.357 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:49.357 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:49.618 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.618 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:49.618 "name": "Existed_Raid", 00:10:49.618 "uuid": "be506031-d795-4948-a20b-118460f481e7", 00:10:49.618 "strip_size_kb": 0, 00:10:49.618 "state": "configuring", 00:10:49.618 "raid_level": "raid1", 00:10:49.618 "superblock": true, 00:10:49.618 "num_base_bdevs": 4, 00:10:49.618 "num_base_bdevs_discovered": 1, 00:10:49.618 "num_base_bdevs_operational": 4, 00:10:49.618 "base_bdevs_list": [ 00:10:49.618 { 00:10:49.618 "name": "BaseBdev1", 00:10:49.618 "uuid": "8f008574-465a-4b23-8961-e5b7228b439f", 00:10:49.618 "is_configured": true, 00:10:49.618 "data_offset": 2048, 00:10:49.618 "data_size": 63488 00:10:49.618 }, 00:10:49.618 { 00:10:49.618 "name": "BaseBdev2", 00:10:49.618 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:49.618 "is_configured": false, 00:10:49.618 "data_offset": 0, 00:10:49.618 "data_size": 0 00:10:49.618 }, 00:10:49.618 { 00:10:49.618 "name": "BaseBdev3", 00:10:49.618 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:49.618 "is_configured": false, 00:10:49.618 "data_offset": 0, 00:10:49.618 "data_size": 0 00:10:49.618 }, 00:10:49.618 { 00:10:49.618 "name": "BaseBdev4", 00:10:49.618 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:49.618 "is_configured": false, 00:10:49.618 "data_offset": 0, 00:10:49.618 "data_size": 0 00:10:49.618 } 00:10:49.618 ] 00:10:49.618 }' 00:10:49.618 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:49.618 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:49.878 [2024-11-27 15:15:17.874129] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:49.878 BaseBdev2 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:49.878 [ 00:10:49.878 { 00:10:49.878 "name": "BaseBdev2", 00:10:49.878 "aliases": [ 00:10:49.878 "fa5a8803-068b-4eb4-9611-6b41683a235a" 00:10:49.878 ], 00:10:49.878 "product_name": "Malloc disk", 00:10:49.878 "block_size": 512, 00:10:49.878 "num_blocks": 65536, 00:10:49.878 "uuid": "fa5a8803-068b-4eb4-9611-6b41683a235a", 00:10:49.878 "assigned_rate_limits": { 00:10:49.878 "rw_ios_per_sec": 0, 00:10:49.878 "rw_mbytes_per_sec": 0, 00:10:49.878 "r_mbytes_per_sec": 0, 00:10:49.878 "w_mbytes_per_sec": 0 00:10:49.878 }, 00:10:49.878 "claimed": true, 00:10:49.878 "claim_type": "exclusive_write", 00:10:49.878 "zoned": false, 00:10:49.878 "supported_io_types": { 00:10:49.878 "read": true, 00:10:49.878 "write": true, 00:10:49.878 "unmap": true, 00:10:49.878 "flush": true, 00:10:49.878 "reset": true, 00:10:49.878 "nvme_admin": false, 00:10:49.878 "nvme_io": false, 00:10:49.878 "nvme_io_md": false, 00:10:49.878 "write_zeroes": true, 00:10:49.878 "zcopy": true, 00:10:49.878 "get_zone_info": false, 00:10:49.878 "zone_management": false, 00:10:49.878 "zone_append": false, 00:10:49.878 "compare": false, 00:10:49.878 "compare_and_write": false, 00:10:49.878 "abort": true, 00:10:49.878 "seek_hole": false, 00:10:49.878 "seek_data": false, 00:10:49.878 "copy": true, 00:10:49.878 "nvme_iov_md": false 00:10:49.878 }, 00:10:49.878 "memory_domains": [ 00:10:49.878 { 00:10:49.878 "dma_device_id": "system", 00:10:49.878 "dma_device_type": 1 00:10:49.878 }, 00:10:49.878 { 00:10:49.878 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:49.878 "dma_device_type": 2 00:10:49.878 } 00:10:49.878 ], 00:10:49.878 "driver_specific": {} 00:10:49.878 } 00:10:49.878 ] 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.878 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:49.878 "name": "Existed_Raid", 00:10:49.878 "uuid": "be506031-d795-4948-a20b-118460f481e7", 00:10:49.878 "strip_size_kb": 0, 00:10:49.878 "state": "configuring", 00:10:49.878 "raid_level": "raid1", 00:10:49.878 "superblock": true, 00:10:49.878 "num_base_bdevs": 4, 00:10:49.878 "num_base_bdevs_discovered": 2, 00:10:49.878 "num_base_bdevs_operational": 4, 00:10:49.878 "base_bdevs_list": [ 00:10:49.878 { 00:10:49.878 "name": "BaseBdev1", 00:10:49.878 "uuid": "8f008574-465a-4b23-8961-e5b7228b439f", 00:10:49.878 "is_configured": true, 00:10:49.878 "data_offset": 2048, 00:10:49.878 "data_size": 63488 00:10:49.878 }, 00:10:49.878 { 00:10:49.878 "name": "BaseBdev2", 00:10:49.878 "uuid": "fa5a8803-068b-4eb4-9611-6b41683a235a", 00:10:49.878 "is_configured": true, 00:10:49.878 "data_offset": 2048, 00:10:49.878 "data_size": 63488 00:10:49.878 }, 00:10:49.878 { 00:10:49.878 "name": "BaseBdev3", 00:10:49.878 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:49.878 "is_configured": false, 00:10:49.878 "data_offset": 0, 00:10:49.878 "data_size": 0 00:10:49.878 }, 00:10:49.878 { 00:10:49.879 "name": "BaseBdev4", 00:10:49.879 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:49.879 "is_configured": false, 00:10:49.879 "data_offset": 0, 00:10:49.879 "data_size": 0 00:10:49.879 } 00:10:49.879 ] 00:10:49.879 }' 00:10:49.879 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:49.879 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:50.448 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:50.448 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:50.448 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:50.448 [2024-11-27 15:15:18.364486] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:50.448 BaseBdev3 00:10:50.448 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:50.448 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:10:50.448 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:50.448 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:50.448 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:50.448 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:50.448 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:50.448 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:50.448 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:50.448 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:50.448 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:50.448 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:50.448 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:50.448 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:50.448 [ 00:10:50.448 { 00:10:50.448 "name": "BaseBdev3", 00:10:50.448 "aliases": [ 00:10:50.448 "ec152b9f-f840-4543-8a22-32f1827170c6" 00:10:50.448 ], 00:10:50.448 "product_name": "Malloc disk", 00:10:50.448 "block_size": 512, 00:10:50.448 "num_blocks": 65536, 00:10:50.448 "uuid": "ec152b9f-f840-4543-8a22-32f1827170c6", 00:10:50.448 "assigned_rate_limits": { 00:10:50.448 "rw_ios_per_sec": 0, 00:10:50.448 "rw_mbytes_per_sec": 0, 00:10:50.448 "r_mbytes_per_sec": 0, 00:10:50.448 "w_mbytes_per_sec": 0 00:10:50.448 }, 00:10:50.448 "claimed": true, 00:10:50.448 "claim_type": "exclusive_write", 00:10:50.448 "zoned": false, 00:10:50.448 "supported_io_types": { 00:10:50.448 "read": true, 00:10:50.448 "write": true, 00:10:50.448 "unmap": true, 00:10:50.448 "flush": true, 00:10:50.448 "reset": true, 00:10:50.448 "nvme_admin": false, 00:10:50.448 "nvme_io": false, 00:10:50.448 "nvme_io_md": false, 00:10:50.448 "write_zeroes": true, 00:10:50.448 "zcopy": true, 00:10:50.448 "get_zone_info": false, 00:10:50.448 "zone_management": false, 00:10:50.448 "zone_append": false, 00:10:50.448 "compare": false, 00:10:50.448 "compare_and_write": false, 00:10:50.448 "abort": true, 00:10:50.448 "seek_hole": false, 00:10:50.449 "seek_data": false, 00:10:50.449 "copy": true, 00:10:50.449 "nvme_iov_md": false 00:10:50.449 }, 00:10:50.449 "memory_domains": [ 00:10:50.449 { 00:10:50.449 "dma_device_id": "system", 00:10:50.449 "dma_device_type": 1 00:10:50.449 }, 00:10:50.449 { 00:10:50.449 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:50.449 "dma_device_type": 2 00:10:50.449 } 00:10:50.449 ], 00:10:50.449 "driver_specific": {} 00:10:50.449 } 00:10:50.449 ] 00:10:50.449 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:50.449 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:50.449 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:50.449 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:50.449 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:50.449 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:50.449 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:50.449 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:50.449 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:50.449 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:50.449 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:50.449 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:50.449 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:50.449 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:50.449 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:50.449 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:50.449 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:50.449 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:50.449 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:50.449 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:50.449 "name": "Existed_Raid", 00:10:50.449 "uuid": "be506031-d795-4948-a20b-118460f481e7", 00:10:50.449 "strip_size_kb": 0, 00:10:50.449 "state": "configuring", 00:10:50.449 "raid_level": "raid1", 00:10:50.449 "superblock": true, 00:10:50.449 "num_base_bdevs": 4, 00:10:50.449 "num_base_bdevs_discovered": 3, 00:10:50.449 "num_base_bdevs_operational": 4, 00:10:50.449 "base_bdevs_list": [ 00:10:50.449 { 00:10:50.449 "name": "BaseBdev1", 00:10:50.449 "uuid": "8f008574-465a-4b23-8961-e5b7228b439f", 00:10:50.449 "is_configured": true, 00:10:50.449 "data_offset": 2048, 00:10:50.449 "data_size": 63488 00:10:50.449 }, 00:10:50.449 { 00:10:50.449 "name": "BaseBdev2", 00:10:50.449 "uuid": "fa5a8803-068b-4eb4-9611-6b41683a235a", 00:10:50.449 "is_configured": true, 00:10:50.449 "data_offset": 2048, 00:10:50.449 "data_size": 63488 00:10:50.449 }, 00:10:50.449 { 00:10:50.449 "name": "BaseBdev3", 00:10:50.449 "uuid": "ec152b9f-f840-4543-8a22-32f1827170c6", 00:10:50.449 "is_configured": true, 00:10:50.449 "data_offset": 2048, 00:10:50.449 "data_size": 63488 00:10:50.449 }, 00:10:50.449 { 00:10:50.449 "name": "BaseBdev4", 00:10:50.449 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:50.449 "is_configured": false, 00:10:50.449 "data_offset": 0, 00:10:50.449 "data_size": 0 00:10:50.449 } 00:10:50.449 ] 00:10:50.449 }' 00:10:50.449 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:50.449 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:51.019 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:51.019 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:51.019 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:51.019 [2024-11-27 15:15:18.834814] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:51.019 [2024-11-27 15:15:18.835150] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:10:51.019 [2024-11-27 15:15:18.835171] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:51.019 [2024-11-27 15:15:18.835433] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:10:51.019 BaseBdev4 00:10:51.019 [2024-11-27 15:15:18.835599] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:10:51.019 [2024-11-27 15:15:18.835613] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:10:51.019 [2024-11-27 15:15:18.835736] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:51.019 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:51.019 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:10:51.019 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:10:51.019 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:51.019 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:51.019 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:51.019 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:51.019 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:51.019 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:51.019 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:51.019 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:51.019 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:51.019 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:51.019 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:51.019 [ 00:10:51.019 { 00:10:51.019 "name": "BaseBdev4", 00:10:51.019 "aliases": [ 00:10:51.019 "27a34ed0-b3b6-4030-8b4a-650928688f75" 00:10:51.019 ], 00:10:51.019 "product_name": "Malloc disk", 00:10:51.019 "block_size": 512, 00:10:51.019 "num_blocks": 65536, 00:10:51.019 "uuid": "27a34ed0-b3b6-4030-8b4a-650928688f75", 00:10:51.019 "assigned_rate_limits": { 00:10:51.019 "rw_ios_per_sec": 0, 00:10:51.019 "rw_mbytes_per_sec": 0, 00:10:51.019 "r_mbytes_per_sec": 0, 00:10:51.019 "w_mbytes_per_sec": 0 00:10:51.019 }, 00:10:51.019 "claimed": true, 00:10:51.019 "claim_type": "exclusive_write", 00:10:51.019 "zoned": false, 00:10:51.019 "supported_io_types": { 00:10:51.019 "read": true, 00:10:51.019 "write": true, 00:10:51.019 "unmap": true, 00:10:51.019 "flush": true, 00:10:51.020 "reset": true, 00:10:51.020 "nvme_admin": false, 00:10:51.020 "nvme_io": false, 00:10:51.020 "nvme_io_md": false, 00:10:51.020 "write_zeroes": true, 00:10:51.020 "zcopy": true, 00:10:51.020 "get_zone_info": false, 00:10:51.020 "zone_management": false, 00:10:51.020 "zone_append": false, 00:10:51.020 "compare": false, 00:10:51.020 "compare_and_write": false, 00:10:51.020 "abort": true, 00:10:51.020 "seek_hole": false, 00:10:51.020 "seek_data": false, 00:10:51.020 "copy": true, 00:10:51.020 "nvme_iov_md": false 00:10:51.020 }, 00:10:51.020 "memory_domains": [ 00:10:51.020 { 00:10:51.020 "dma_device_id": "system", 00:10:51.020 "dma_device_type": 1 00:10:51.020 }, 00:10:51.020 { 00:10:51.020 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:51.020 "dma_device_type": 2 00:10:51.020 } 00:10:51.020 ], 00:10:51.020 "driver_specific": {} 00:10:51.020 } 00:10:51.020 ] 00:10:51.020 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:51.020 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:51.020 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:51.020 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:51.020 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:10:51.020 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:51.020 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:51.020 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:51.020 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:51.020 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:51.020 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:51.020 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:51.020 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:51.020 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:51.020 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:51.020 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:51.020 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:51.020 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:51.020 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:51.020 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:51.020 "name": "Existed_Raid", 00:10:51.020 "uuid": "be506031-d795-4948-a20b-118460f481e7", 00:10:51.020 "strip_size_kb": 0, 00:10:51.020 "state": "online", 00:10:51.020 "raid_level": "raid1", 00:10:51.020 "superblock": true, 00:10:51.020 "num_base_bdevs": 4, 00:10:51.020 "num_base_bdevs_discovered": 4, 00:10:51.020 "num_base_bdevs_operational": 4, 00:10:51.020 "base_bdevs_list": [ 00:10:51.020 { 00:10:51.020 "name": "BaseBdev1", 00:10:51.020 "uuid": "8f008574-465a-4b23-8961-e5b7228b439f", 00:10:51.020 "is_configured": true, 00:10:51.020 "data_offset": 2048, 00:10:51.020 "data_size": 63488 00:10:51.020 }, 00:10:51.020 { 00:10:51.020 "name": "BaseBdev2", 00:10:51.020 "uuid": "fa5a8803-068b-4eb4-9611-6b41683a235a", 00:10:51.020 "is_configured": true, 00:10:51.020 "data_offset": 2048, 00:10:51.020 "data_size": 63488 00:10:51.020 }, 00:10:51.020 { 00:10:51.020 "name": "BaseBdev3", 00:10:51.020 "uuid": "ec152b9f-f840-4543-8a22-32f1827170c6", 00:10:51.020 "is_configured": true, 00:10:51.020 "data_offset": 2048, 00:10:51.020 "data_size": 63488 00:10:51.020 }, 00:10:51.020 { 00:10:51.020 "name": "BaseBdev4", 00:10:51.020 "uuid": "27a34ed0-b3b6-4030-8b4a-650928688f75", 00:10:51.020 "is_configured": true, 00:10:51.020 "data_offset": 2048, 00:10:51.020 "data_size": 63488 00:10:51.020 } 00:10:51.020 ] 00:10:51.020 }' 00:10:51.020 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:51.020 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:51.283 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:51.283 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:51.283 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:51.283 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:51.283 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:51.283 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:51.283 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:51.283 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:51.283 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:51.283 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:51.283 [2024-11-27 15:15:19.298395] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:51.283 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:51.283 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:51.283 "name": "Existed_Raid", 00:10:51.283 "aliases": [ 00:10:51.283 "be506031-d795-4948-a20b-118460f481e7" 00:10:51.283 ], 00:10:51.283 "product_name": "Raid Volume", 00:10:51.283 "block_size": 512, 00:10:51.283 "num_blocks": 63488, 00:10:51.283 "uuid": "be506031-d795-4948-a20b-118460f481e7", 00:10:51.283 "assigned_rate_limits": { 00:10:51.283 "rw_ios_per_sec": 0, 00:10:51.283 "rw_mbytes_per_sec": 0, 00:10:51.283 "r_mbytes_per_sec": 0, 00:10:51.283 "w_mbytes_per_sec": 0 00:10:51.283 }, 00:10:51.283 "claimed": false, 00:10:51.283 "zoned": false, 00:10:51.283 "supported_io_types": { 00:10:51.283 "read": true, 00:10:51.283 "write": true, 00:10:51.283 "unmap": false, 00:10:51.283 "flush": false, 00:10:51.283 "reset": true, 00:10:51.283 "nvme_admin": false, 00:10:51.283 "nvme_io": false, 00:10:51.283 "nvme_io_md": false, 00:10:51.283 "write_zeroes": true, 00:10:51.283 "zcopy": false, 00:10:51.283 "get_zone_info": false, 00:10:51.283 "zone_management": false, 00:10:51.283 "zone_append": false, 00:10:51.283 "compare": false, 00:10:51.283 "compare_and_write": false, 00:10:51.283 "abort": false, 00:10:51.283 "seek_hole": false, 00:10:51.283 "seek_data": false, 00:10:51.283 "copy": false, 00:10:51.283 "nvme_iov_md": false 00:10:51.283 }, 00:10:51.283 "memory_domains": [ 00:10:51.283 { 00:10:51.283 "dma_device_id": "system", 00:10:51.283 "dma_device_type": 1 00:10:51.283 }, 00:10:51.283 { 00:10:51.283 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:51.283 "dma_device_type": 2 00:10:51.283 }, 00:10:51.283 { 00:10:51.283 "dma_device_id": "system", 00:10:51.283 "dma_device_type": 1 00:10:51.283 }, 00:10:51.283 { 00:10:51.283 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:51.283 "dma_device_type": 2 00:10:51.283 }, 00:10:51.283 { 00:10:51.283 "dma_device_id": "system", 00:10:51.283 "dma_device_type": 1 00:10:51.283 }, 00:10:51.283 { 00:10:51.283 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:51.283 "dma_device_type": 2 00:10:51.283 }, 00:10:51.283 { 00:10:51.283 "dma_device_id": "system", 00:10:51.283 "dma_device_type": 1 00:10:51.283 }, 00:10:51.283 { 00:10:51.283 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:51.283 "dma_device_type": 2 00:10:51.283 } 00:10:51.283 ], 00:10:51.283 "driver_specific": { 00:10:51.283 "raid": { 00:10:51.283 "uuid": "be506031-d795-4948-a20b-118460f481e7", 00:10:51.283 "strip_size_kb": 0, 00:10:51.283 "state": "online", 00:10:51.283 "raid_level": "raid1", 00:10:51.283 "superblock": true, 00:10:51.283 "num_base_bdevs": 4, 00:10:51.283 "num_base_bdevs_discovered": 4, 00:10:51.283 "num_base_bdevs_operational": 4, 00:10:51.283 "base_bdevs_list": [ 00:10:51.283 { 00:10:51.283 "name": "BaseBdev1", 00:10:51.283 "uuid": "8f008574-465a-4b23-8961-e5b7228b439f", 00:10:51.283 "is_configured": true, 00:10:51.283 "data_offset": 2048, 00:10:51.283 "data_size": 63488 00:10:51.283 }, 00:10:51.283 { 00:10:51.283 "name": "BaseBdev2", 00:10:51.283 "uuid": "fa5a8803-068b-4eb4-9611-6b41683a235a", 00:10:51.283 "is_configured": true, 00:10:51.283 "data_offset": 2048, 00:10:51.284 "data_size": 63488 00:10:51.284 }, 00:10:51.284 { 00:10:51.284 "name": "BaseBdev3", 00:10:51.284 "uuid": "ec152b9f-f840-4543-8a22-32f1827170c6", 00:10:51.284 "is_configured": true, 00:10:51.284 "data_offset": 2048, 00:10:51.284 "data_size": 63488 00:10:51.284 }, 00:10:51.284 { 00:10:51.284 "name": "BaseBdev4", 00:10:51.284 "uuid": "27a34ed0-b3b6-4030-8b4a-650928688f75", 00:10:51.284 "is_configured": true, 00:10:51.284 "data_offset": 2048, 00:10:51.284 "data_size": 63488 00:10:51.284 } 00:10:51.284 ] 00:10:51.284 } 00:10:51.284 } 00:10:51.284 }' 00:10:51.284 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:51.284 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:51.284 BaseBdev2 00:10:51.284 BaseBdev3 00:10:51.284 BaseBdev4' 00:10:51.284 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:51.543 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:51.543 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:51.543 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:51.543 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:51.543 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:51.543 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:51.543 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:51.543 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:51.543 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:51.543 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:51.543 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:51.544 [2024-11-27 15:15:19.593615] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:51.544 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:51.803 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:51.803 "name": "Existed_Raid", 00:10:51.803 "uuid": "be506031-d795-4948-a20b-118460f481e7", 00:10:51.803 "strip_size_kb": 0, 00:10:51.803 "state": "online", 00:10:51.803 "raid_level": "raid1", 00:10:51.803 "superblock": true, 00:10:51.803 "num_base_bdevs": 4, 00:10:51.803 "num_base_bdevs_discovered": 3, 00:10:51.803 "num_base_bdevs_operational": 3, 00:10:51.803 "base_bdevs_list": [ 00:10:51.803 { 00:10:51.803 "name": null, 00:10:51.803 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:51.803 "is_configured": false, 00:10:51.803 "data_offset": 0, 00:10:51.803 "data_size": 63488 00:10:51.803 }, 00:10:51.803 { 00:10:51.803 "name": "BaseBdev2", 00:10:51.803 "uuid": "fa5a8803-068b-4eb4-9611-6b41683a235a", 00:10:51.803 "is_configured": true, 00:10:51.803 "data_offset": 2048, 00:10:51.803 "data_size": 63488 00:10:51.803 }, 00:10:51.803 { 00:10:51.803 "name": "BaseBdev3", 00:10:51.803 "uuid": "ec152b9f-f840-4543-8a22-32f1827170c6", 00:10:51.803 "is_configured": true, 00:10:51.803 "data_offset": 2048, 00:10:51.803 "data_size": 63488 00:10:51.803 }, 00:10:51.803 { 00:10:51.803 "name": "BaseBdev4", 00:10:51.803 "uuid": "27a34ed0-b3b6-4030-8b4a-650928688f75", 00:10:51.803 "is_configured": true, 00:10:51.803 "data_offset": 2048, 00:10:51.803 "data_size": 63488 00:10:51.803 } 00:10:51.803 ] 00:10:51.803 }' 00:10:51.803 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:51.804 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:52.064 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:52.064 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:52.064 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:52.064 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.064 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:52.064 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:52.064 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.064 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:52.064 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:52.064 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:52.064 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.064 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:52.064 [2024-11-27 15:15:20.096307] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:52.064 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.064 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:52.064 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:52.064 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:52.064 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:52.064 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.064 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:52.064 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.064 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:52.064 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:52.064 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:52.064 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.064 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:52.064 [2024-11-27 15:15:20.155431] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:52.325 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.325 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:52.325 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:52.325 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:52.325 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.325 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:52.326 [2024-11-27 15:15:20.226474] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:10:52.326 [2024-11-27 15:15:20.226622] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:52.326 [2024-11-27 15:15:20.238079] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:52.326 [2024-11-27 15:15:20.238193] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:52.326 [2024-11-27 15:15:20.238233] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:52.326 BaseBdev2 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:52.326 [ 00:10:52.326 { 00:10:52.326 "name": "BaseBdev2", 00:10:52.326 "aliases": [ 00:10:52.326 "8092cd16-9cb8-4930-bfd3-6ad1e5b00f57" 00:10:52.326 ], 00:10:52.326 "product_name": "Malloc disk", 00:10:52.326 "block_size": 512, 00:10:52.326 "num_blocks": 65536, 00:10:52.326 "uuid": "8092cd16-9cb8-4930-bfd3-6ad1e5b00f57", 00:10:52.326 "assigned_rate_limits": { 00:10:52.326 "rw_ios_per_sec": 0, 00:10:52.326 "rw_mbytes_per_sec": 0, 00:10:52.326 "r_mbytes_per_sec": 0, 00:10:52.326 "w_mbytes_per_sec": 0 00:10:52.326 }, 00:10:52.326 "claimed": false, 00:10:52.326 "zoned": false, 00:10:52.326 "supported_io_types": { 00:10:52.326 "read": true, 00:10:52.326 "write": true, 00:10:52.326 "unmap": true, 00:10:52.326 "flush": true, 00:10:52.326 "reset": true, 00:10:52.326 "nvme_admin": false, 00:10:52.326 "nvme_io": false, 00:10:52.326 "nvme_io_md": false, 00:10:52.326 "write_zeroes": true, 00:10:52.326 "zcopy": true, 00:10:52.326 "get_zone_info": false, 00:10:52.326 "zone_management": false, 00:10:52.326 "zone_append": false, 00:10:52.326 "compare": false, 00:10:52.326 "compare_and_write": false, 00:10:52.326 "abort": true, 00:10:52.326 "seek_hole": false, 00:10:52.326 "seek_data": false, 00:10:52.326 "copy": true, 00:10:52.326 "nvme_iov_md": false 00:10:52.326 }, 00:10:52.326 "memory_domains": [ 00:10:52.326 { 00:10:52.326 "dma_device_id": "system", 00:10:52.326 "dma_device_type": 1 00:10:52.326 }, 00:10:52.326 { 00:10:52.326 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:52.326 "dma_device_type": 2 00:10:52.326 } 00:10:52.326 ], 00:10:52.326 "driver_specific": {} 00:10:52.326 } 00:10:52.326 ] 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:52.326 BaseBdev3 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.326 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:52.326 [ 00:10:52.326 { 00:10:52.326 "name": "BaseBdev3", 00:10:52.326 "aliases": [ 00:10:52.326 "0eb506ad-4bcb-4041-949f-18a1ca66f11b" 00:10:52.326 ], 00:10:52.326 "product_name": "Malloc disk", 00:10:52.326 "block_size": 512, 00:10:52.326 "num_blocks": 65536, 00:10:52.326 "uuid": "0eb506ad-4bcb-4041-949f-18a1ca66f11b", 00:10:52.326 "assigned_rate_limits": { 00:10:52.326 "rw_ios_per_sec": 0, 00:10:52.326 "rw_mbytes_per_sec": 0, 00:10:52.326 "r_mbytes_per_sec": 0, 00:10:52.326 "w_mbytes_per_sec": 0 00:10:52.326 }, 00:10:52.326 "claimed": false, 00:10:52.326 "zoned": false, 00:10:52.326 "supported_io_types": { 00:10:52.326 "read": true, 00:10:52.326 "write": true, 00:10:52.326 "unmap": true, 00:10:52.326 "flush": true, 00:10:52.326 "reset": true, 00:10:52.326 "nvme_admin": false, 00:10:52.326 "nvme_io": false, 00:10:52.326 "nvme_io_md": false, 00:10:52.326 "write_zeroes": true, 00:10:52.326 "zcopy": true, 00:10:52.326 "get_zone_info": false, 00:10:52.326 "zone_management": false, 00:10:52.326 "zone_append": false, 00:10:52.326 "compare": false, 00:10:52.326 "compare_and_write": false, 00:10:52.326 "abort": true, 00:10:52.326 "seek_hole": false, 00:10:52.326 "seek_data": false, 00:10:52.326 "copy": true, 00:10:52.327 "nvme_iov_md": false 00:10:52.327 }, 00:10:52.327 "memory_domains": [ 00:10:52.327 { 00:10:52.327 "dma_device_id": "system", 00:10:52.327 "dma_device_type": 1 00:10:52.327 }, 00:10:52.327 { 00:10:52.327 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:52.327 "dma_device_type": 2 00:10:52.327 } 00:10:52.327 ], 00:10:52.327 "driver_specific": {} 00:10:52.327 } 00:10:52.327 ] 00:10:52.327 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.327 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:52.327 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:52.327 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:52.327 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:52.327 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.327 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:52.327 BaseBdev4 00:10:52.327 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.327 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:10:52.327 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:10:52.327 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:52.327 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:52.327 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:52.327 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:52.327 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:52.327 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.327 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:52.327 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.327 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:52.327 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.327 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:52.587 [ 00:10:52.587 { 00:10:52.587 "name": "BaseBdev4", 00:10:52.587 "aliases": [ 00:10:52.587 "00b814c5-6d7c-46ce-807a-034c28cc27f5" 00:10:52.587 ], 00:10:52.587 "product_name": "Malloc disk", 00:10:52.587 "block_size": 512, 00:10:52.587 "num_blocks": 65536, 00:10:52.587 "uuid": "00b814c5-6d7c-46ce-807a-034c28cc27f5", 00:10:52.587 "assigned_rate_limits": { 00:10:52.587 "rw_ios_per_sec": 0, 00:10:52.587 "rw_mbytes_per_sec": 0, 00:10:52.587 "r_mbytes_per_sec": 0, 00:10:52.587 "w_mbytes_per_sec": 0 00:10:52.587 }, 00:10:52.587 "claimed": false, 00:10:52.587 "zoned": false, 00:10:52.587 "supported_io_types": { 00:10:52.587 "read": true, 00:10:52.587 "write": true, 00:10:52.587 "unmap": true, 00:10:52.587 "flush": true, 00:10:52.587 "reset": true, 00:10:52.587 "nvme_admin": false, 00:10:52.587 "nvme_io": false, 00:10:52.587 "nvme_io_md": false, 00:10:52.587 "write_zeroes": true, 00:10:52.587 "zcopy": true, 00:10:52.587 "get_zone_info": false, 00:10:52.587 "zone_management": false, 00:10:52.587 "zone_append": false, 00:10:52.587 "compare": false, 00:10:52.587 "compare_and_write": false, 00:10:52.587 "abort": true, 00:10:52.587 "seek_hole": false, 00:10:52.587 "seek_data": false, 00:10:52.587 "copy": true, 00:10:52.587 "nvme_iov_md": false 00:10:52.587 }, 00:10:52.587 "memory_domains": [ 00:10:52.587 { 00:10:52.587 "dma_device_id": "system", 00:10:52.587 "dma_device_type": 1 00:10:52.587 }, 00:10:52.587 { 00:10:52.587 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:52.587 "dma_device_type": 2 00:10:52.587 } 00:10:52.587 ], 00:10:52.587 "driver_specific": {} 00:10:52.587 } 00:10:52.587 ] 00:10:52.587 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.587 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:52.587 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:52.587 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:52.587 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:52.587 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.587 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:52.587 [2024-11-27 15:15:20.454035] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:52.587 [2024-11-27 15:15:20.454095] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:52.587 [2024-11-27 15:15:20.454115] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:52.587 [2024-11-27 15:15:20.455916] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:52.587 [2024-11-27 15:15:20.455978] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:52.587 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.587 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:52.587 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:52.587 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:52.587 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:52.587 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:52.587 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:52.587 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:52.587 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:52.587 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:52.587 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:52.587 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:52.587 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.587 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:52.587 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:52.587 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.587 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:52.587 "name": "Existed_Raid", 00:10:52.587 "uuid": "6e507082-b7b0-4ea6-bf4b-7f5d8750a598", 00:10:52.587 "strip_size_kb": 0, 00:10:52.587 "state": "configuring", 00:10:52.587 "raid_level": "raid1", 00:10:52.587 "superblock": true, 00:10:52.587 "num_base_bdevs": 4, 00:10:52.587 "num_base_bdevs_discovered": 3, 00:10:52.587 "num_base_bdevs_operational": 4, 00:10:52.587 "base_bdevs_list": [ 00:10:52.587 { 00:10:52.587 "name": "BaseBdev1", 00:10:52.587 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:52.587 "is_configured": false, 00:10:52.587 "data_offset": 0, 00:10:52.587 "data_size": 0 00:10:52.587 }, 00:10:52.587 { 00:10:52.587 "name": "BaseBdev2", 00:10:52.587 "uuid": "8092cd16-9cb8-4930-bfd3-6ad1e5b00f57", 00:10:52.587 "is_configured": true, 00:10:52.587 "data_offset": 2048, 00:10:52.587 "data_size": 63488 00:10:52.587 }, 00:10:52.587 { 00:10:52.587 "name": "BaseBdev3", 00:10:52.587 "uuid": "0eb506ad-4bcb-4041-949f-18a1ca66f11b", 00:10:52.587 "is_configured": true, 00:10:52.587 "data_offset": 2048, 00:10:52.587 "data_size": 63488 00:10:52.587 }, 00:10:52.587 { 00:10:52.588 "name": "BaseBdev4", 00:10:52.588 "uuid": "00b814c5-6d7c-46ce-807a-034c28cc27f5", 00:10:52.588 "is_configured": true, 00:10:52.588 "data_offset": 2048, 00:10:52.588 "data_size": 63488 00:10:52.588 } 00:10:52.588 ] 00:10:52.588 }' 00:10:52.588 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:52.588 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:52.847 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:52.847 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.847 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:52.847 [2024-11-27 15:15:20.861345] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:52.847 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.847 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:52.847 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:52.847 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:52.847 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:52.847 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:52.847 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:52.847 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:52.847 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:52.847 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:52.847 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:52.847 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:52.847 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.847 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:52.847 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:52.847 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.847 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:52.847 "name": "Existed_Raid", 00:10:52.847 "uuid": "6e507082-b7b0-4ea6-bf4b-7f5d8750a598", 00:10:52.847 "strip_size_kb": 0, 00:10:52.847 "state": "configuring", 00:10:52.847 "raid_level": "raid1", 00:10:52.847 "superblock": true, 00:10:52.847 "num_base_bdevs": 4, 00:10:52.847 "num_base_bdevs_discovered": 2, 00:10:52.847 "num_base_bdevs_operational": 4, 00:10:52.847 "base_bdevs_list": [ 00:10:52.847 { 00:10:52.847 "name": "BaseBdev1", 00:10:52.847 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:52.847 "is_configured": false, 00:10:52.847 "data_offset": 0, 00:10:52.847 "data_size": 0 00:10:52.847 }, 00:10:52.847 { 00:10:52.847 "name": null, 00:10:52.847 "uuid": "8092cd16-9cb8-4930-bfd3-6ad1e5b00f57", 00:10:52.847 "is_configured": false, 00:10:52.847 "data_offset": 0, 00:10:52.847 "data_size": 63488 00:10:52.847 }, 00:10:52.847 { 00:10:52.847 "name": "BaseBdev3", 00:10:52.847 "uuid": "0eb506ad-4bcb-4041-949f-18a1ca66f11b", 00:10:52.847 "is_configured": true, 00:10:52.847 "data_offset": 2048, 00:10:52.847 "data_size": 63488 00:10:52.847 }, 00:10:52.847 { 00:10:52.847 "name": "BaseBdev4", 00:10:52.847 "uuid": "00b814c5-6d7c-46ce-807a-034c28cc27f5", 00:10:52.847 "is_configured": true, 00:10:52.847 "data_offset": 2048, 00:10:52.847 "data_size": 63488 00:10:52.847 } 00:10:52.847 ] 00:10:52.847 }' 00:10:52.847 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:52.847 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:53.416 [2024-11-27 15:15:21.355568] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:53.416 BaseBdev1 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:53.416 [ 00:10:53.416 { 00:10:53.416 "name": "BaseBdev1", 00:10:53.416 "aliases": [ 00:10:53.416 "824a204b-bbe9-4770-b534-81e214f97d9e" 00:10:53.416 ], 00:10:53.416 "product_name": "Malloc disk", 00:10:53.416 "block_size": 512, 00:10:53.416 "num_blocks": 65536, 00:10:53.416 "uuid": "824a204b-bbe9-4770-b534-81e214f97d9e", 00:10:53.416 "assigned_rate_limits": { 00:10:53.416 "rw_ios_per_sec": 0, 00:10:53.416 "rw_mbytes_per_sec": 0, 00:10:53.416 "r_mbytes_per_sec": 0, 00:10:53.416 "w_mbytes_per_sec": 0 00:10:53.416 }, 00:10:53.416 "claimed": true, 00:10:53.416 "claim_type": "exclusive_write", 00:10:53.416 "zoned": false, 00:10:53.416 "supported_io_types": { 00:10:53.416 "read": true, 00:10:53.416 "write": true, 00:10:53.416 "unmap": true, 00:10:53.416 "flush": true, 00:10:53.416 "reset": true, 00:10:53.416 "nvme_admin": false, 00:10:53.416 "nvme_io": false, 00:10:53.416 "nvme_io_md": false, 00:10:53.416 "write_zeroes": true, 00:10:53.416 "zcopy": true, 00:10:53.416 "get_zone_info": false, 00:10:53.416 "zone_management": false, 00:10:53.416 "zone_append": false, 00:10:53.416 "compare": false, 00:10:53.416 "compare_and_write": false, 00:10:53.416 "abort": true, 00:10:53.416 "seek_hole": false, 00:10:53.416 "seek_data": false, 00:10:53.416 "copy": true, 00:10:53.416 "nvme_iov_md": false 00:10:53.416 }, 00:10:53.416 "memory_domains": [ 00:10:53.416 { 00:10:53.416 "dma_device_id": "system", 00:10:53.416 "dma_device_type": 1 00:10:53.416 }, 00:10:53.416 { 00:10:53.416 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:53.416 "dma_device_type": 2 00:10:53.416 } 00:10:53.416 ], 00:10:53.416 "driver_specific": {} 00:10:53.416 } 00:10:53.416 ] 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:53.416 "name": "Existed_Raid", 00:10:53.416 "uuid": "6e507082-b7b0-4ea6-bf4b-7f5d8750a598", 00:10:53.416 "strip_size_kb": 0, 00:10:53.416 "state": "configuring", 00:10:53.416 "raid_level": "raid1", 00:10:53.416 "superblock": true, 00:10:53.416 "num_base_bdevs": 4, 00:10:53.416 "num_base_bdevs_discovered": 3, 00:10:53.416 "num_base_bdevs_operational": 4, 00:10:53.416 "base_bdevs_list": [ 00:10:53.416 { 00:10:53.416 "name": "BaseBdev1", 00:10:53.416 "uuid": "824a204b-bbe9-4770-b534-81e214f97d9e", 00:10:53.416 "is_configured": true, 00:10:53.416 "data_offset": 2048, 00:10:53.416 "data_size": 63488 00:10:53.416 }, 00:10:53.416 { 00:10:53.416 "name": null, 00:10:53.416 "uuid": "8092cd16-9cb8-4930-bfd3-6ad1e5b00f57", 00:10:53.416 "is_configured": false, 00:10:53.416 "data_offset": 0, 00:10:53.416 "data_size": 63488 00:10:53.416 }, 00:10:53.416 { 00:10:53.416 "name": "BaseBdev3", 00:10:53.416 "uuid": "0eb506ad-4bcb-4041-949f-18a1ca66f11b", 00:10:53.416 "is_configured": true, 00:10:53.416 "data_offset": 2048, 00:10:53.416 "data_size": 63488 00:10:53.416 }, 00:10:53.416 { 00:10:53.416 "name": "BaseBdev4", 00:10:53.416 "uuid": "00b814c5-6d7c-46ce-807a-034c28cc27f5", 00:10:53.416 "is_configured": true, 00:10:53.416 "data_offset": 2048, 00:10:53.416 "data_size": 63488 00:10:53.416 } 00:10:53.416 ] 00:10:53.416 }' 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:53.416 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:53.984 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:53.984 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.984 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:53.984 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:53.984 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.984 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:53.984 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:53.984 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.984 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:53.984 [2024-11-27 15:15:21.902687] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:53.984 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.984 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:53.984 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:53.984 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:53.984 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:53.984 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:53.984 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:53.984 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:53.984 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:53.984 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:53.984 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:53.984 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:53.984 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:53.984 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.984 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:53.984 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.984 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:53.984 "name": "Existed_Raid", 00:10:53.984 "uuid": "6e507082-b7b0-4ea6-bf4b-7f5d8750a598", 00:10:53.984 "strip_size_kb": 0, 00:10:53.984 "state": "configuring", 00:10:53.984 "raid_level": "raid1", 00:10:53.984 "superblock": true, 00:10:53.984 "num_base_bdevs": 4, 00:10:53.984 "num_base_bdevs_discovered": 2, 00:10:53.984 "num_base_bdevs_operational": 4, 00:10:53.984 "base_bdevs_list": [ 00:10:53.984 { 00:10:53.984 "name": "BaseBdev1", 00:10:53.984 "uuid": "824a204b-bbe9-4770-b534-81e214f97d9e", 00:10:53.984 "is_configured": true, 00:10:53.984 "data_offset": 2048, 00:10:53.984 "data_size": 63488 00:10:53.984 }, 00:10:53.984 { 00:10:53.984 "name": null, 00:10:53.984 "uuid": "8092cd16-9cb8-4930-bfd3-6ad1e5b00f57", 00:10:53.984 "is_configured": false, 00:10:53.984 "data_offset": 0, 00:10:53.984 "data_size": 63488 00:10:53.984 }, 00:10:53.984 { 00:10:53.984 "name": null, 00:10:53.984 "uuid": "0eb506ad-4bcb-4041-949f-18a1ca66f11b", 00:10:53.984 "is_configured": false, 00:10:53.984 "data_offset": 0, 00:10:53.984 "data_size": 63488 00:10:53.984 }, 00:10:53.984 { 00:10:53.984 "name": "BaseBdev4", 00:10:53.984 "uuid": "00b814c5-6d7c-46ce-807a-034c28cc27f5", 00:10:53.984 "is_configured": true, 00:10:53.984 "data_offset": 2048, 00:10:53.984 "data_size": 63488 00:10:53.984 } 00:10:53.984 ] 00:10:53.984 }' 00:10:53.984 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:53.984 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:54.243 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:54.243 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.243 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:54.243 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:54.243 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.503 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:54.503 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:54.503 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.503 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:54.503 [2024-11-27 15:15:22.353964] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:54.503 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.503 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:54.503 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:54.503 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:54.503 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:54.503 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:54.503 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:54.503 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:54.503 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:54.503 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:54.503 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:54.503 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:54.503 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.503 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:54.503 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:54.503 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.503 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:54.503 "name": "Existed_Raid", 00:10:54.503 "uuid": "6e507082-b7b0-4ea6-bf4b-7f5d8750a598", 00:10:54.503 "strip_size_kb": 0, 00:10:54.503 "state": "configuring", 00:10:54.503 "raid_level": "raid1", 00:10:54.503 "superblock": true, 00:10:54.503 "num_base_bdevs": 4, 00:10:54.503 "num_base_bdevs_discovered": 3, 00:10:54.503 "num_base_bdevs_operational": 4, 00:10:54.503 "base_bdevs_list": [ 00:10:54.503 { 00:10:54.503 "name": "BaseBdev1", 00:10:54.503 "uuid": "824a204b-bbe9-4770-b534-81e214f97d9e", 00:10:54.503 "is_configured": true, 00:10:54.503 "data_offset": 2048, 00:10:54.503 "data_size": 63488 00:10:54.503 }, 00:10:54.503 { 00:10:54.503 "name": null, 00:10:54.503 "uuid": "8092cd16-9cb8-4930-bfd3-6ad1e5b00f57", 00:10:54.503 "is_configured": false, 00:10:54.503 "data_offset": 0, 00:10:54.503 "data_size": 63488 00:10:54.503 }, 00:10:54.503 { 00:10:54.503 "name": "BaseBdev3", 00:10:54.503 "uuid": "0eb506ad-4bcb-4041-949f-18a1ca66f11b", 00:10:54.503 "is_configured": true, 00:10:54.503 "data_offset": 2048, 00:10:54.503 "data_size": 63488 00:10:54.503 }, 00:10:54.503 { 00:10:54.503 "name": "BaseBdev4", 00:10:54.503 "uuid": "00b814c5-6d7c-46ce-807a-034c28cc27f5", 00:10:54.503 "is_configured": true, 00:10:54.503 "data_offset": 2048, 00:10:54.503 "data_size": 63488 00:10:54.503 } 00:10:54.503 ] 00:10:54.503 }' 00:10:54.503 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:54.503 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:54.763 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:54.763 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:54.763 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.763 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:54.763 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.763 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:54.763 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:54.763 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.763 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:54.763 [2024-11-27 15:15:22.829128] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:54.763 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.763 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:54.763 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:54.763 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:54.763 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:54.763 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:54.763 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:54.763 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:54.763 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:54.763 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:54.763 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:54.763 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:54.763 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:54.763 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.763 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:55.022 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.022 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:55.022 "name": "Existed_Raid", 00:10:55.022 "uuid": "6e507082-b7b0-4ea6-bf4b-7f5d8750a598", 00:10:55.022 "strip_size_kb": 0, 00:10:55.022 "state": "configuring", 00:10:55.022 "raid_level": "raid1", 00:10:55.022 "superblock": true, 00:10:55.022 "num_base_bdevs": 4, 00:10:55.022 "num_base_bdevs_discovered": 2, 00:10:55.022 "num_base_bdevs_operational": 4, 00:10:55.022 "base_bdevs_list": [ 00:10:55.022 { 00:10:55.022 "name": null, 00:10:55.022 "uuid": "824a204b-bbe9-4770-b534-81e214f97d9e", 00:10:55.022 "is_configured": false, 00:10:55.022 "data_offset": 0, 00:10:55.022 "data_size": 63488 00:10:55.022 }, 00:10:55.022 { 00:10:55.022 "name": null, 00:10:55.022 "uuid": "8092cd16-9cb8-4930-bfd3-6ad1e5b00f57", 00:10:55.022 "is_configured": false, 00:10:55.022 "data_offset": 0, 00:10:55.022 "data_size": 63488 00:10:55.022 }, 00:10:55.022 { 00:10:55.022 "name": "BaseBdev3", 00:10:55.022 "uuid": "0eb506ad-4bcb-4041-949f-18a1ca66f11b", 00:10:55.022 "is_configured": true, 00:10:55.022 "data_offset": 2048, 00:10:55.022 "data_size": 63488 00:10:55.022 }, 00:10:55.022 { 00:10:55.022 "name": "BaseBdev4", 00:10:55.022 "uuid": "00b814c5-6d7c-46ce-807a-034c28cc27f5", 00:10:55.022 "is_configured": true, 00:10:55.022 "data_offset": 2048, 00:10:55.022 "data_size": 63488 00:10:55.022 } 00:10:55.023 ] 00:10:55.023 }' 00:10:55.023 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:55.023 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:55.282 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:55.282 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:55.282 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.282 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:55.282 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.282 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:55.282 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:55.282 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.282 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:55.282 [2024-11-27 15:15:23.350786] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:55.282 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.282 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:55.282 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:55.282 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:55.282 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:55.282 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:55.282 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:55.282 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:55.282 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:55.282 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:55.282 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:55.282 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:55.282 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:55.282 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.282 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:55.282 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.542 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:55.542 "name": "Existed_Raid", 00:10:55.542 "uuid": "6e507082-b7b0-4ea6-bf4b-7f5d8750a598", 00:10:55.542 "strip_size_kb": 0, 00:10:55.542 "state": "configuring", 00:10:55.542 "raid_level": "raid1", 00:10:55.542 "superblock": true, 00:10:55.542 "num_base_bdevs": 4, 00:10:55.542 "num_base_bdevs_discovered": 3, 00:10:55.542 "num_base_bdevs_operational": 4, 00:10:55.542 "base_bdevs_list": [ 00:10:55.542 { 00:10:55.542 "name": null, 00:10:55.542 "uuid": "824a204b-bbe9-4770-b534-81e214f97d9e", 00:10:55.542 "is_configured": false, 00:10:55.542 "data_offset": 0, 00:10:55.542 "data_size": 63488 00:10:55.542 }, 00:10:55.542 { 00:10:55.542 "name": "BaseBdev2", 00:10:55.542 "uuid": "8092cd16-9cb8-4930-bfd3-6ad1e5b00f57", 00:10:55.542 "is_configured": true, 00:10:55.542 "data_offset": 2048, 00:10:55.542 "data_size": 63488 00:10:55.542 }, 00:10:55.542 { 00:10:55.542 "name": "BaseBdev3", 00:10:55.542 "uuid": "0eb506ad-4bcb-4041-949f-18a1ca66f11b", 00:10:55.542 "is_configured": true, 00:10:55.542 "data_offset": 2048, 00:10:55.542 "data_size": 63488 00:10:55.542 }, 00:10:55.542 { 00:10:55.542 "name": "BaseBdev4", 00:10:55.542 "uuid": "00b814c5-6d7c-46ce-807a-034c28cc27f5", 00:10:55.542 "is_configured": true, 00:10:55.542 "data_offset": 2048, 00:10:55.542 "data_size": 63488 00:10:55.542 } 00:10:55.542 ] 00:10:55.542 }' 00:10:55.542 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:55.542 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:55.802 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:55.802 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:55.802 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.802 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:55.802 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.802 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:55.802 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:55.802 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:55.802 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.802 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:55.802 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.802 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 824a204b-bbe9-4770-b534-81e214f97d9e 00:10:55.802 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.802 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:55.802 [2024-11-27 15:15:23.861061] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:55.802 [2024-11-27 15:15:23.861345] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:10:55.802 [2024-11-27 15:15:23.861387] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:55.802 [2024-11-27 15:15:23.861686] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:10:55.802 NewBaseBdev 00:10:55.802 [2024-11-27 15:15:23.861852] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:10:55.802 [2024-11-27 15:15:23.861865] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006d00 00:10:55.802 [2024-11-27 15:15:23.861980] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:55.803 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.803 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:55.803 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:10:55.803 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:55.803 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:55.803 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:55.803 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:55.803 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:55.803 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.803 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:55.803 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.803 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:55.803 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.803 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:55.803 [ 00:10:55.803 { 00:10:55.803 "name": "NewBaseBdev", 00:10:55.803 "aliases": [ 00:10:55.803 "824a204b-bbe9-4770-b534-81e214f97d9e" 00:10:55.803 ], 00:10:55.803 "product_name": "Malloc disk", 00:10:55.803 "block_size": 512, 00:10:55.803 "num_blocks": 65536, 00:10:55.803 "uuid": "824a204b-bbe9-4770-b534-81e214f97d9e", 00:10:55.803 "assigned_rate_limits": { 00:10:55.803 "rw_ios_per_sec": 0, 00:10:55.803 "rw_mbytes_per_sec": 0, 00:10:55.803 "r_mbytes_per_sec": 0, 00:10:55.803 "w_mbytes_per_sec": 0 00:10:55.803 }, 00:10:55.803 "claimed": true, 00:10:55.803 "claim_type": "exclusive_write", 00:10:55.803 "zoned": false, 00:10:55.803 "supported_io_types": { 00:10:55.803 "read": true, 00:10:55.803 "write": true, 00:10:55.803 "unmap": true, 00:10:55.803 "flush": true, 00:10:55.803 "reset": true, 00:10:55.803 "nvme_admin": false, 00:10:55.803 "nvme_io": false, 00:10:55.803 "nvme_io_md": false, 00:10:55.803 "write_zeroes": true, 00:10:55.803 "zcopy": true, 00:10:55.803 "get_zone_info": false, 00:10:55.803 "zone_management": false, 00:10:55.803 "zone_append": false, 00:10:55.803 "compare": false, 00:10:55.803 "compare_and_write": false, 00:10:55.803 "abort": true, 00:10:55.803 "seek_hole": false, 00:10:55.803 "seek_data": false, 00:10:55.803 "copy": true, 00:10:55.803 "nvme_iov_md": false 00:10:55.803 }, 00:10:55.803 "memory_domains": [ 00:10:55.803 { 00:10:55.803 "dma_device_id": "system", 00:10:55.803 "dma_device_type": 1 00:10:55.803 }, 00:10:55.803 { 00:10:55.803 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:55.803 "dma_device_type": 2 00:10:55.803 } 00:10:55.803 ], 00:10:55.803 "driver_specific": {} 00:10:55.803 } 00:10:55.803 ] 00:10:55.803 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.803 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:55.803 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:10:55.803 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:55.803 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:55.803 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:55.803 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:55.803 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:55.803 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:55.803 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:55.803 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:55.803 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:55.803 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:56.062 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.062 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:56.062 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:56.062 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.062 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:56.062 "name": "Existed_Raid", 00:10:56.062 "uuid": "6e507082-b7b0-4ea6-bf4b-7f5d8750a598", 00:10:56.062 "strip_size_kb": 0, 00:10:56.062 "state": "online", 00:10:56.062 "raid_level": "raid1", 00:10:56.062 "superblock": true, 00:10:56.062 "num_base_bdevs": 4, 00:10:56.062 "num_base_bdevs_discovered": 4, 00:10:56.062 "num_base_bdevs_operational": 4, 00:10:56.062 "base_bdevs_list": [ 00:10:56.062 { 00:10:56.062 "name": "NewBaseBdev", 00:10:56.062 "uuid": "824a204b-bbe9-4770-b534-81e214f97d9e", 00:10:56.062 "is_configured": true, 00:10:56.062 "data_offset": 2048, 00:10:56.062 "data_size": 63488 00:10:56.062 }, 00:10:56.062 { 00:10:56.062 "name": "BaseBdev2", 00:10:56.062 "uuid": "8092cd16-9cb8-4930-bfd3-6ad1e5b00f57", 00:10:56.062 "is_configured": true, 00:10:56.062 "data_offset": 2048, 00:10:56.063 "data_size": 63488 00:10:56.063 }, 00:10:56.063 { 00:10:56.063 "name": "BaseBdev3", 00:10:56.063 "uuid": "0eb506ad-4bcb-4041-949f-18a1ca66f11b", 00:10:56.063 "is_configured": true, 00:10:56.063 "data_offset": 2048, 00:10:56.063 "data_size": 63488 00:10:56.063 }, 00:10:56.063 { 00:10:56.063 "name": "BaseBdev4", 00:10:56.063 "uuid": "00b814c5-6d7c-46ce-807a-034c28cc27f5", 00:10:56.063 "is_configured": true, 00:10:56.063 "data_offset": 2048, 00:10:56.063 "data_size": 63488 00:10:56.063 } 00:10:56.063 ] 00:10:56.063 }' 00:10:56.063 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:56.063 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:56.322 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:56.322 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:56.322 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:56.322 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:56.322 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:56.322 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:56.322 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:56.322 15:15:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.322 15:15:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:56.322 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:56.322 [2024-11-27 15:15:24.376595] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:56.322 15:15:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.322 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:56.322 "name": "Existed_Raid", 00:10:56.322 "aliases": [ 00:10:56.322 "6e507082-b7b0-4ea6-bf4b-7f5d8750a598" 00:10:56.322 ], 00:10:56.322 "product_name": "Raid Volume", 00:10:56.322 "block_size": 512, 00:10:56.322 "num_blocks": 63488, 00:10:56.322 "uuid": "6e507082-b7b0-4ea6-bf4b-7f5d8750a598", 00:10:56.322 "assigned_rate_limits": { 00:10:56.322 "rw_ios_per_sec": 0, 00:10:56.322 "rw_mbytes_per_sec": 0, 00:10:56.322 "r_mbytes_per_sec": 0, 00:10:56.322 "w_mbytes_per_sec": 0 00:10:56.322 }, 00:10:56.322 "claimed": false, 00:10:56.322 "zoned": false, 00:10:56.322 "supported_io_types": { 00:10:56.322 "read": true, 00:10:56.322 "write": true, 00:10:56.322 "unmap": false, 00:10:56.322 "flush": false, 00:10:56.322 "reset": true, 00:10:56.322 "nvme_admin": false, 00:10:56.322 "nvme_io": false, 00:10:56.322 "nvme_io_md": false, 00:10:56.322 "write_zeroes": true, 00:10:56.322 "zcopy": false, 00:10:56.322 "get_zone_info": false, 00:10:56.322 "zone_management": false, 00:10:56.322 "zone_append": false, 00:10:56.322 "compare": false, 00:10:56.322 "compare_and_write": false, 00:10:56.322 "abort": false, 00:10:56.322 "seek_hole": false, 00:10:56.322 "seek_data": false, 00:10:56.322 "copy": false, 00:10:56.322 "nvme_iov_md": false 00:10:56.322 }, 00:10:56.322 "memory_domains": [ 00:10:56.322 { 00:10:56.322 "dma_device_id": "system", 00:10:56.322 "dma_device_type": 1 00:10:56.322 }, 00:10:56.322 { 00:10:56.322 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:56.322 "dma_device_type": 2 00:10:56.322 }, 00:10:56.322 { 00:10:56.322 "dma_device_id": "system", 00:10:56.322 "dma_device_type": 1 00:10:56.322 }, 00:10:56.322 { 00:10:56.322 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:56.322 "dma_device_type": 2 00:10:56.322 }, 00:10:56.322 { 00:10:56.322 "dma_device_id": "system", 00:10:56.322 "dma_device_type": 1 00:10:56.322 }, 00:10:56.322 { 00:10:56.322 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:56.322 "dma_device_type": 2 00:10:56.322 }, 00:10:56.322 { 00:10:56.322 "dma_device_id": "system", 00:10:56.322 "dma_device_type": 1 00:10:56.322 }, 00:10:56.322 { 00:10:56.322 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:56.322 "dma_device_type": 2 00:10:56.322 } 00:10:56.322 ], 00:10:56.322 "driver_specific": { 00:10:56.322 "raid": { 00:10:56.322 "uuid": "6e507082-b7b0-4ea6-bf4b-7f5d8750a598", 00:10:56.322 "strip_size_kb": 0, 00:10:56.322 "state": "online", 00:10:56.322 "raid_level": "raid1", 00:10:56.322 "superblock": true, 00:10:56.322 "num_base_bdevs": 4, 00:10:56.322 "num_base_bdevs_discovered": 4, 00:10:56.322 "num_base_bdevs_operational": 4, 00:10:56.322 "base_bdevs_list": [ 00:10:56.322 { 00:10:56.322 "name": "NewBaseBdev", 00:10:56.322 "uuid": "824a204b-bbe9-4770-b534-81e214f97d9e", 00:10:56.322 "is_configured": true, 00:10:56.322 "data_offset": 2048, 00:10:56.322 "data_size": 63488 00:10:56.322 }, 00:10:56.322 { 00:10:56.322 "name": "BaseBdev2", 00:10:56.322 "uuid": "8092cd16-9cb8-4930-bfd3-6ad1e5b00f57", 00:10:56.322 "is_configured": true, 00:10:56.322 "data_offset": 2048, 00:10:56.322 "data_size": 63488 00:10:56.322 }, 00:10:56.322 { 00:10:56.322 "name": "BaseBdev3", 00:10:56.322 "uuid": "0eb506ad-4bcb-4041-949f-18a1ca66f11b", 00:10:56.322 "is_configured": true, 00:10:56.322 "data_offset": 2048, 00:10:56.322 "data_size": 63488 00:10:56.322 }, 00:10:56.323 { 00:10:56.323 "name": "BaseBdev4", 00:10:56.323 "uuid": "00b814c5-6d7c-46ce-807a-034c28cc27f5", 00:10:56.323 "is_configured": true, 00:10:56.323 "data_offset": 2048, 00:10:56.323 "data_size": 63488 00:10:56.323 } 00:10:56.323 ] 00:10:56.323 } 00:10:56.323 } 00:10:56.323 }' 00:10:56.323 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:56.582 BaseBdev2 00:10:56.582 BaseBdev3 00:10:56.582 BaseBdev4' 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.582 15:15:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:56.842 [2024-11-27 15:15:24.691772] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:56.842 [2024-11-27 15:15:24.691847] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:56.842 [2024-11-27 15:15:24.691949] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:56.842 [2024-11-27 15:15:24.692216] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:56.842 [2024-11-27 15:15:24.692273] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name Existed_Raid, state offline 00:10:56.842 15:15:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.842 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 84772 00:10:56.842 15:15:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 84772 ']' 00:10:56.842 15:15:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 84772 00:10:56.842 15:15:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:10:56.842 15:15:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:56.842 15:15:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 84772 00:10:56.842 killing process with pid 84772 00:10:56.842 15:15:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:56.842 15:15:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:56.842 15:15:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 84772' 00:10:56.842 15:15:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 84772 00:10:56.842 [2024-11-27 15:15:24.740016] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:56.842 15:15:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 84772 00:10:56.842 [2024-11-27 15:15:24.781046] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:57.102 15:15:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:10:57.102 00:10:57.103 real 0m9.454s 00:10:57.103 user 0m16.038s 00:10:57.103 sys 0m2.086s 00:10:57.103 15:15:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:57.103 15:15:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:57.103 ************************************ 00:10:57.103 END TEST raid_state_function_test_sb 00:10:57.103 ************************************ 00:10:57.103 15:15:25 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 4 00:10:57.103 15:15:25 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:10:57.103 15:15:25 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:57.103 15:15:25 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:57.103 ************************************ 00:10:57.103 START TEST raid_superblock_test 00:10:57.103 ************************************ 00:10:57.103 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 4 00:10:57.103 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:10:57.103 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:10:57.103 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:10:57.103 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:10:57.103 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:10:57.103 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:10:57.103 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:10:57.103 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:10:57.103 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:10:57.103 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:10:57.103 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:10:57.103 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:10:57.103 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:10:57.103 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:10:57.103 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:10:57.103 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=85420 00:10:57.103 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:10:57.103 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 85420 00:10:57.103 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 85420 ']' 00:10:57.103 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:57.103 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:57.103 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:57.103 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:57.103 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:57.103 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.103 [2024-11-27 15:15:25.162327] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:10:57.103 [2024-11-27 15:15:25.162528] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid85420 ] 00:10:57.362 [2024-11-27 15:15:25.335450] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:57.362 [2024-11-27 15:15:25.360987] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:57.362 [2024-11-27 15:15:25.404257] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:57.362 [2024-11-27 15:15:25.404392] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:57.932 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:57.932 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:10:57.932 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:10:57.932 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:57.932 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:10:57.932 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:10:57.932 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:10:57.932 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:57.932 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:57.932 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:57.932 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:10:57.932 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:57.932 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.932 malloc1 00:10:57.932 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:57.932 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:57.932 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:57.932 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.932 [2024-11-27 15:15:26.028483] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:57.932 [2024-11-27 15:15:26.028647] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:57.932 [2024-11-27 15:15:26.028687] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:10:57.932 [2024-11-27 15:15:26.028721] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:57.932 [2024-11-27 15:15:26.030964] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:57.932 [2024-11-27 15:15:26.031044] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:57.932 pt1 00:10:57.932 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:57.932 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:57.932 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:57.932 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:10:57.932 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:10:57.932 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:10:57.932 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:57.932 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.192 malloc2 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.192 [2024-11-27 15:15:26.057314] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:58.192 [2024-11-27 15:15:26.057399] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:58.192 [2024-11-27 15:15:26.057416] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:10:58.192 [2024-11-27 15:15:26.057427] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:58.192 [2024-11-27 15:15:26.059642] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:58.192 [2024-11-27 15:15:26.059715] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:58.192 pt2 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.192 malloc3 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.192 [2024-11-27 15:15:26.090054] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:58.192 [2024-11-27 15:15:26.090187] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:58.192 [2024-11-27 15:15:26.090249] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:10:58.192 [2024-11-27 15:15:26.090286] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:58.192 [2024-11-27 15:15:26.092352] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:58.192 [2024-11-27 15:15:26.092429] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:58.192 pt3 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.192 malloc4 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.192 [2024-11-27 15:15:26.130162] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:58.192 [2024-11-27 15:15:26.130296] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:58.192 [2024-11-27 15:15:26.130331] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:58.192 [2024-11-27 15:15:26.130364] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:58.192 [2024-11-27 15:15:26.132454] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:58.192 [2024-11-27 15:15:26.132530] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:58.192 pt4 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.192 [2024-11-27 15:15:26.142172] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:58.192 [2024-11-27 15:15:26.143981] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:58.192 [2024-11-27 15:15:26.144041] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:58.192 [2024-11-27 15:15:26.144107] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:58.192 [2024-11-27 15:15:26.144259] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:10:58.192 [2024-11-27 15:15:26.144272] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:58.192 [2024-11-27 15:15:26.144539] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:10:58.192 [2024-11-27 15:15:26.144690] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:10:58.192 [2024-11-27 15:15:26.144700] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:10:58.192 [2024-11-27 15:15:26.144816] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:58.192 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:58.193 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:58.193 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:58.193 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:58.193 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:58.193 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:58.193 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:58.193 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:58.193 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:58.193 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.193 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.193 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.193 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:58.193 "name": "raid_bdev1", 00:10:58.193 "uuid": "04fb4881-7d28-4de8-97e1-80b94e75d55b", 00:10:58.193 "strip_size_kb": 0, 00:10:58.193 "state": "online", 00:10:58.193 "raid_level": "raid1", 00:10:58.193 "superblock": true, 00:10:58.193 "num_base_bdevs": 4, 00:10:58.193 "num_base_bdevs_discovered": 4, 00:10:58.193 "num_base_bdevs_operational": 4, 00:10:58.193 "base_bdevs_list": [ 00:10:58.193 { 00:10:58.193 "name": "pt1", 00:10:58.193 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:58.193 "is_configured": true, 00:10:58.193 "data_offset": 2048, 00:10:58.193 "data_size": 63488 00:10:58.193 }, 00:10:58.193 { 00:10:58.193 "name": "pt2", 00:10:58.193 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:58.193 "is_configured": true, 00:10:58.193 "data_offset": 2048, 00:10:58.193 "data_size": 63488 00:10:58.193 }, 00:10:58.193 { 00:10:58.193 "name": "pt3", 00:10:58.193 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:58.193 "is_configured": true, 00:10:58.193 "data_offset": 2048, 00:10:58.193 "data_size": 63488 00:10:58.193 }, 00:10:58.193 { 00:10:58.193 "name": "pt4", 00:10:58.193 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:58.193 "is_configured": true, 00:10:58.193 "data_offset": 2048, 00:10:58.193 "data_size": 63488 00:10:58.193 } 00:10:58.193 ] 00:10:58.193 }' 00:10:58.193 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:58.193 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.452 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:10:58.452 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:58.452 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:58.452 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:58.452 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:58.452 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:58.452 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:58.452 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.452 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.452 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:58.452 [2024-11-27 15:15:26.537836] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:58.713 "name": "raid_bdev1", 00:10:58.713 "aliases": [ 00:10:58.713 "04fb4881-7d28-4de8-97e1-80b94e75d55b" 00:10:58.713 ], 00:10:58.713 "product_name": "Raid Volume", 00:10:58.713 "block_size": 512, 00:10:58.713 "num_blocks": 63488, 00:10:58.713 "uuid": "04fb4881-7d28-4de8-97e1-80b94e75d55b", 00:10:58.713 "assigned_rate_limits": { 00:10:58.713 "rw_ios_per_sec": 0, 00:10:58.713 "rw_mbytes_per_sec": 0, 00:10:58.713 "r_mbytes_per_sec": 0, 00:10:58.713 "w_mbytes_per_sec": 0 00:10:58.713 }, 00:10:58.713 "claimed": false, 00:10:58.713 "zoned": false, 00:10:58.713 "supported_io_types": { 00:10:58.713 "read": true, 00:10:58.713 "write": true, 00:10:58.713 "unmap": false, 00:10:58.713 "flush": false, 00:10:58.713 "reset": true, 00:10:58.713 "nvme_admin": false, 00:10:58.713 "nvme_io": false, 00:10:58.713 "nvme_io_md": false, 00:10:58.713 "write_zeroes": true, 00:10:58.713 "zcopy": false, 00:10:58.713 "get_zone_info": false, 00:10:58.713 "zone_management": false, 00:10:58.713 "zone_append": false, 00:10:58.713 "compare": false, 00:10:58.713 "compare_and_write": false, 00:10:58.713 "abort": false, 00:10:58.713 "seek_hole": false, 00:10:58.713 "seek_data": false, 00:10:58.713 "copy": false, 00:10:58.713 "nvme_iov_md": false 00:10:58.713 }, 00:10:58.713 "memory_domains": [ 00:10:58.713 { 00:10:58.713 "dma_device_id": "system", 00:10:58.713 "dma_device_type": 1 00:10:58.713 }, 00:10:58.713 { 00:10:58.713 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:58.713 "dma_device_type": 2 00:10:58.713 }, 00:10:58.713 { 00:10:58.713 "dma_device_id": "system", 00:10:58.713 "dma_device_type": 1 00:10:58.713 }, 00:10:58.713 { 00:10:58.713 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:58.713 "dma_device_type": 2 00:10:58.713 }, 00:10:58.713 { 00:10:58.713 "dma_device_id": "system", 00:10:58.713 "dma_device_type": 1 00:10:58.713 }, 00:10:58.713 { 00:10:58.713 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:58.713 "dma_device_type": 2 00:10:58.713 }, 00:10:58.713 { 00:10:58.713 "dma_device_id": "system", 00:10:58.713 "dma_device_type": 1 00:10:58.713 }, 00:10:58.713 { 00:10:58.713 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:58.713 "dma_device_type": 2 00:10:58.713 } 00:10:58.713 ], 00:10:58.713 "driver_specific": { 00:10:58.713 "raid": { 00:10:58.713 "uuid": "04fb4881-7d28-4de8-97e1-80b94e75d55b", 00:10:58.713 "strip_size_kb": 0, 00:10:58.713 "state": "online", 00:10:58.713 "raid_level": "raid1", 00:10:58.713 "superblock": true, 00:10:58.713 "num_base_bdevs": 4, 00:10:58.713 "num_base_bdevs_discovered": 4, 00:10:58.713 "num_base_bdevs_operational": 4, 00:10:58.713 "base_bdevs_list": [ 00:10:58.713 { 00:10:58.713 "name": "pt1", 00:10:58.713 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:58.713 "is_configured": true, 00:10:58.713 "data_offset": 2048, 00:10:58.713 "data_size": 63488 00:10:58.713 }, 00:10:58.713 { 00:10:58.713 "name": "pt2", 00:10:58.713 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:58.713 "is_configured": true, 00:10:58.713 "data_offset": 2048, 00:10:58.713 "data_size": 63488 00:10:58.713 }, 00:10:58.713 { 00:10:58.713 "name": "pt3", 00:10:58.713 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:58.713 "is_configured": true, 00:10:58.713 "data_offset": 2048, 00:10:58.713 "data_size": 63488 00:10:58.713 }, 00:10:58.713 { 00:10:58.713 "name": "pt4", 00:10:58.713 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:58.713 "is_configured": true, 00:10:58.713 "data_offset": 2048, 00:10:58.713 "data_size": 63488 00:10:58.713 } 00:10:58.713 ] 00:10:58.713 } 00:10:58.713 } 00:10:58.713 }' 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:58.713 pt2 00:10:58.713 pt3 00:10:58.713 pt4' 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.713 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.974 [2024-11-27 15:15:26.841275] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=04fb4881-7d28-4de8-97e1-80b94e75d55b 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 04fb4881-7d28-4de8-97e1-80b94e75d55b ']' 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.974 [2024-11-27 15:15:26.884881] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:58.974 [2024-11-27 15:15:26.884923] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:58.974 [2024-11-27 15:15:26.885004] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:58.974 [2024-11-27 15:15:26.885089] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:58.974 [2024-11-27 15:15:26.885107] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.974 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.974 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.974 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:10:58.974 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:58.974 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:10:58.974 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:58.974 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:10:58.974 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:10:58.974 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:10:58.974 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:10:58.974 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:58.974 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.974 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.974 [2024-11-27 15:15:27.052652] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:10:58.974 [2024-11-27 15:15:27.054537] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:10:58.974 [2024-11-27 15:15:27.054632] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:10:58.974 [2024-11-27 15:15:27.054677] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:10:58.974 [2024-11-27 15:15:27.054749] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:10:58.974 [2024-11-27 15:15:27.054836] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:10:58.974 [2024-11-27 15:15:27.054860] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:10:58.975 [2024-11-27 15:15:27.054877] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:10:58.975 [2024-11-27 15:15:27.054891] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:58.975 [2024-11-27 15:15:27.054913] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state configuring 00:10:58.975 request: 00:10:58.975 { 00:10:58.975 "name": "raid_bdev1", 00:10:58.975 "raid_level": "raid1", 00:10:58.975 "base_bdevs": [ 00:10:58.975 "malloc1", 00:10:58.975 "malloc2", 00:10:58.975 "malloc3", 00:10:58.975 "malloc4" 00:10:58.975 ], 00:10:58.975 "superblock": false, 00:10:58.975 "method": "bdev_raid_create", 00:10:58.975 "req_id": 1 00:10:58.975 } 00:10:58.975 Got JSON-RPC error response 00:10:58.975 response: 00:10:58.975 { 00:10:58.975 "code": -17, 00:10:58.975 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:10:58.975 } 00:10:58.975 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:10:58.975 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:10:58.975 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:10:58.975 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:10:58.975 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:10:58.975 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:58.975 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.975 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.975 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:10:58.975 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:59.234 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:10:59.234 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:10:59.234 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:59.234 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:59.234 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:59.234 [2024-11-27 15:15:27.120449] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:59.234 [2024-11-27 15:15:27.120548] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:59.234 [2024-11-27 15:15:27.120586] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:10:59.234 [2024-11-27 15:15:27.120613] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:59.234 [2024-11-27 15:15:27.122693] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:59.234 [2024-11-27 15:15:27.122761] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:59.234 [2024-11-27 15:15:27.122868] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:10:59.234 [2024-11-27 15:15:27.122944] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:59.234 pt1 00:10:59.234 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:59.234 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 4 00:10:59.234 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:59.234 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:59.234 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:59.234 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:59.234 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:59.234 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:59.234 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:59.234 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:59.234 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:59.234 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:59.234 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:59.234 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:59.234 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:59.234 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:59.235 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:59.235 "name": "raid_bdev1", 00:10:59.235 "uuid": "04fb4881-7d28-4de8-97e1-80b94e75d55b", 00:10:59.235 "strip_size_kb": 0, 00:10:59.235 "state": "configuring", 00:10:59.235 "raid_level": "raid1", 00:10:59.235 "superblock": true, 00:10:59.235 "num_base_bdevs": 4, 00:10:59.235 "num_base_bdevs_discovered": 1, 00:10:59.235 "num_base_bdevs_operational": 4, 00:10:59.235 "base_bdevs_list": [ 00:10:59.235 { 00:10:59.235 "name": "pt1", 00:10:59.235 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:59.235 "is_configured": true, 00:10:59.235 "data_offset": 2048, 00:10:59.235 "data_size": 63488 00:10:59.235 }, 00:10:59.235 { 00:10:59.235 "name": null, 00:10:59.235 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:59.235 "is_configured": false, 00:10:59.235 "data_offset": 2048, 00:10:59.235 "data_size": 63488 00:10:59.235 }, 00:10:59.235 { 00:10:59.235 "name": null, 00:10:59.235 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:59.235 "is_configured": false, 00:10:59.235 "data_offset": 2048, 00:10:59.235 "data_size": 63488 00:10:59.235 }, 00:10:59.235 { 00:10:59.235 "name": null, 00:10:59.235 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:59.235 "is_configured": false, 00:10:59.235 "data_offset": 2048, 00:10:59.235 "data_size": 63488 00:10:59.235 } 00:10:59.235 ] 00:10:59.235 }' 00:10:59.235 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:59.235 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:59.494 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:10:59.494 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:59.494 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:59.494 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:59.494 [2024-11-27 15:15:27.527804] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:59.494 [2024-11-27 15:15:27.527994] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:59.494 [2024-11-27 15:15:27.528024] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:10:59.494 [2024-11-27 15:15:27.528034] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:59.494 [2024-11-27 15:15:27.528428] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:59.494 [2024-11-27 15:15:27.528445] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:59.494 [2024-11-27 15:15:27.528525] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:59.494 [2024-11-27 15:15:27.528547] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:59.494 pt2 00:10:59.494 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:59.494 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:10:59.494 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:59.494 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:59.494 [2024-11-27 15:15:27.535772] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:10:59.494 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:59.494 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 4 00:10:59.494 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:59.494 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:59.494 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:59.494 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:59.494 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:59.494 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:59.494 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:59.494 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:59.494 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:59.494 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:59.494 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:59.494 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:59.494 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:59.494 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:59.494 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:59.494 "name": "raid_bdev1", 00:10:59.494 "uuid": "04fb4881-7d28-4de8-97e1-80b94e75d55b", 00:10:59.494 "strip_size_kb": 0, 00:10:59.494 "state": "configuring", 00:10:59.494 "raid_level": "raid1", 00:10:59.494 "superblock": true, 00:10:59.494 "num_base_bdevs": 4, 00:10:59.494 "num_base_bdevs_discovered": 1, 00:10:59.494 "num_base_bdevs_operational": 4, 00:10:59.494 "base_bdevs_list": [ 00:10:59.494 { 00:10:59.494 "name": "pt1", 00:10:59.494 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:59.494 "is_configured": true, 00:10:59.494 "data_offset": 2048, 00:10:59.494 "data_size": 63488 00:10:59.494 }, 00:10:59.494 { 00:10:59.494 "name": null, 00:10:59.494 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:59.494 "is_configured": false, 00:10:59.494 "data_offset": 0, 00:10:59.494 "data_size": 63488 00:10:59.494 }, 00:10:59.494 { 00:10:59.494 "name": null, 00:10:59.494 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:59.494 "is_configured": false, 00:10:59.494 "data_offset": 2048, 00:10:59.494 "data_size": 63488 00:10:59.494 }, 00:10:59.494 { 00:10:59.494 "name": null, 00:10:59.494 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:59.494 "is_configured": false, 00:10:59.494 "data_offset": 2048, 00:10:59.494 "data_size": 63488 00:10:59.494 } 00:10:59.494 ] 00:10:59.494 }' 00:10:59.494 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:59.494 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.064 [2024-11-27 15:15:28.007015] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:11:00.064 [2024-11-27 15:15:28.007176] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:00.064 [2024-11-27 15:15:28.007205] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:11:00.064 [2024-11-27 15:15:28.007219] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:00.064 [2024-11-27 15:15:28.007647] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:00.064 [2024-11-27 15:15:28.007670] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:11:00.064 [2024-11-27 15:15:28.007746] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:11:00.064 [2024-11-27 15:15:28.007771] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:11:00.064 pt2 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.064 [2024-11-27 15:15:28.018960] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:11:00.064 [2024-11-27 15:15:28.019016] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:00.064 [2024-11-27 15:15:28.019032] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:11:00.064 [2024-11-27 15:15:28.019042] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:00.064 [2024-11-27 15:15:28.019364] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:00.064 [2024-11-27 15:15:28.019382] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:11:00.064 [2024-11-27 15:15:28.019441] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:11:00.064 [2024-11-27 15:15:28.019463] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:11:00.064 pt3 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.064 [2024-11-27 15:15:28.030915] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:11:00.064 [2024-11-27 15:15:28.030964] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:00.064 [2024-11-27 15:15:28.030976] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:11:00.064 [2024-11-27 15:15:28.030985] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:00.064 [2024-11-27 15:15:28.031268] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:00.064 [2024-11-27 15:15:28.031295] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:11:00.064 [2024-11-27 15:15:28.031347] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:11:00.064 [2024-11-27 15:15:28.031373] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:11:00.064 [2024-11-27 15:15:28.031479] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:11:00.064 [2024-11-27 15:15:28.031489] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:00.064 [2024-11-27 15:15:28.031732] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:11:00.064 [2024-11-27 15:15:28.031855] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:11:00.064 [2024-11-27 15:15:28.031865] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:11:00.064 [2024-11-27 15:15:28.031983] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:00.064 pt4 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.064 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:00.064 "name": "raid_bdev1", 00:11:00.064 "uuid": "04fb4881-7d28-4de8-97e1-80b94e75d55b", 00:11:00.064 "strip_size_kb": 0, 00:11:00.064 "state": "online", 00:11:00.064 "raid_level": "raid1", 00:11:00.064 "superblock": true, 00:11:00.064 "num_base_bdevs": 4, 00:11:00.064 "num_base_bdevs_discovered": 4, 00:11:00.064 "num_base_bdevs_operational": 4, 00:11:00.064 "base_bdevs_list": [ 00:11:00.064 { 00:11:00.064 "name": "pt1", 00:11:00.065 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:00.065 "is_configured": true, 00:11:00.065 "data_offset": 2048, 00:11:00.065 "data_size": 63488 00:11:00.065 }, 00:11:00.065 { 00:11:00.065 "name": "pt2", 00:11:00.065 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:00.065 "is_configured": true, 00:11:00.065 "data_offset": 2048, 00:11:00.065 "data_size": 63488 00:11:00.065 }, 00:11:00.065 { 00:11:00.065 "name": "pt3", 00:11:00.065 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:00.065 "is_configured": true, 00:11:00.065 "data_offset": 2048, 00:11:00.065 "data_size": 63488 00:11:00.065 }, 00:11:00.065 { 00:11:00.065 "name": "pt4", 00:11:00.065 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:00.065 "is_configured": true, 00:11:00.065 "data_offset": 2048, 00:11:00.065 "data_size": 63488 00:11:00.065 } 00:11:00.065 ] 00:11:00.065 }' 00:11:00.065 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:00.065 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.661 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:11:00.661 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:11:00.661 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:00.661 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:00.661 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:11:00.661 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:00.661 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:00.661 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:00.661 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.661 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.661 [2024-11-27 15:15:28.466532] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:00.661 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.661 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:00.661 "name": "raid_bdev1", 00:11:00.661 "aliases": [ 00:11:00.661 "04fb4881-7d28-4de8-97e1-80b94e75d55b" 00:11:00.661 ], 00:11:00.661 "product_name": "Raid Volume", 00:11:00.661 "block_size": 512, 00:11:00.661 "num_blocks": 63488, 00:11:00.661 "uuid": "04fb4881-7d28-4de8-97e1-80b94e75d55b", 00:11:00.661 "assigned_rate_limits": { 00:11:00.661 "rw_ios_per_sec": 0, 00:11:00.661 "rw_mbytes_per_sec": 0, 00:11:00.661 "r_mbytes_per_sec": 0, 00:11:00.661 "w_mbytes_per_sec": 0 00:11:00.661 }, 00:11:00.661 "claimed": false, 00:11:00.661 "zoned": false, 00:11:00.661 "supported_io_types": { 00:11:00.661 "read": true, 00:11:00.661 "write": true, 00:11:00.661 "unmap": false, 00:11:00.661 "flush": false, 00:11:00.661 "reset": true, 00:11:00.661 "nvme_admin": false, 00:11:00.661 "nvme_io": false, 00:11:00.661 "nvme_io_md": false, 00:11:00.661 "write_zeroes": true, 00:11:00.661 "zcopy": false, 00:11:00.661 "get_zone_info": false, 00:11:00.661 "zone_management": false, 00:11:00.661 "zone_append": false, 00:11:00.661 "compare": false, 00:11:00.661 "compare_and_write": false, 00:11:00.661 "abort": false, 00:11:00.661 "seek_hole": false, 00:11:00.661 "seek_data": false, 00:11:00.661 "copy": false, 00:11:00.661 "nvme_iov_md": false 00:11:00.661 }, 00:11:00.661 "memory_domains": [ 00:11:00.661 { 00:11:00.661 "dma_device_id": "system", 00:11:00.661 "dma_device_type": 1 00:11:00.661 }, 00:11:00.661 { 00:11:00.661 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:00.661 "dma_device_type": 2 00:11:00.661 }, 00:11:00.661 { 00:11:00.661 "dma_device_id": "system", 00:11:00.661 "dma_device_type": 1 00:11:00.661 }, 00:11:00.661 { 00:11:00.661 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:00.661 "dma_device_type": 2 00:11:00.661 }, 00:11:00.661 { 00:11:00.661 "dma_device_id": "system", 00:11:00.661 "dma_device_type": 1 00:11:00.661 }, 00:11:00.661 { 00:11:00.661 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:00.661 "dma_device_type": 2 00:11:00.662 }, 00:11:00.662 { 00:11:00.662 "dma_device_id": "system", 00:11:00.662 "dma_device_type": 1 00:11:00.662 }, 00:11:00.662 { 00:11:00.662 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:00.662 "dma_device_type": 2 00:11:00.662 } 00:11:00.662 ], 00:11:00.662 "driver_specific": { 00:11:00.662 "raid": { 00:11:00.662 "uuid": "04fb4881-7d28-4de8-97e1-80b94e75d55b", 00:11:00.662 "strip_size_kb": 0, 00:11:00.662 "state": "online", 00:11:00.662 "raid_level": "raid1", 00:11:00.662 "superblock": true, 00:11:00.662 "num_base_bdevs": 4, 00:11:00.662 "num_base_bdevs_discovered": 4, 00:11:00.662 "num_base_bdevs_operational": 4, 00:11:00.662 "base_bdevs_list": [ 00:11:00.662 { 00:11:00.662 "name": "pt1", 00:11:00.662 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:00.662 "is_configured": true, 00:11:00.662 "data_offset": 2048, 00:11:00.662 "data_size": 63488 00:11:00.662 }, 00:11:00.662 { 00:11:00.662 "name": "pt2", 00:11:00.662 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:00.662 "is_configured": true, 00:11:00.662 "data_offset": 2048, 00:11:00.662 "data_size": 63488 00:11:00.662 }, 00:11:00.662 { 00:11:00.662 "name": "pt3", 00:11:00.662 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:00.662 "is_configured": true, 00:11:00.662 "data_offset": 2048, 00:11:00.662 "data_size": 63488 00:11:00.662 }, 00:11:00.662 { 00:11:00.662 "name": "pt4", 00:11:00.662 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:00.662 "is_configured": true, 00:11:00.662 "data_offset": 2048, 00:11:00.662 "data_size": 63488 00:11:00.662 } 00:11:00.662 ] 00:11:00.662 } 00:11:00.662 } 00:11:00.662 }' 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:11:00.662 pt2 00:11:00.662 pt3 00:11:00.662 pt4' 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.662 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.662 [2024-11-27 15:15:28.750013] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:00.922 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.922 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 04fb4881-7d28-4de8-97e1-80b94e75d55b '!=' 04fb4881-7d28-4de8-97e1-80b94e75d55b ']' 00:11:00.922 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:11:00.922 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:00.922 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:11:00.922 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:11:00.922 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.922 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.922 [2024-11-27 15:15:28.781634] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:11:00.922 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.922 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:11:00.922 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:00.922 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:00.922 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:00.922 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:00.922 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:00.922 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:00.922 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:00.922 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:00.922 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:00.922 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:00.922 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.922 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.922 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:00.922 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.922 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:00.922 "name": "raid_bdev1", 00:11:00.922 "uuid": "04fb4881-7d28-4de8-97e1-80b94e75d55b", 00:11:00.922 "strip_size_kb": 0, 00:11:00.922 "state": "online", 00:11:00.922 "raid_level": "raid1", 00:11:00.922 "superblock": true, 00:11:00.922 "num_base_bdevs": 4, 00:11:00.922 "num_base_bdevs_discovered": 3, 00:11:00.922 "num_base_bdevs_operational": 3, 00:11:00.922 "base_bdevs_list": [ 00:11:00.922 { 00:11:00.922 "name": null, 00:11:00.922 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:00.922 "is_configured": false, 00:11:00.922 "data_offset": 0, 00:11:00.922 "data_size": 63488 00:11:00.922 }, 00:11:00.922 { 00:11:00.922 "name": "pt2", 00:11:00.922 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:00.922 "is_configured": true, 00:11:00.922 "data_offset": 2048, 00:11:00.922 "data_size": 63488 00:11:00.922 }, 00:11:00.922 { 00:11:00.922 "name": "pt3", 00:11:00.922 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:00.922 "is_configured": true, 00:11:00.922 "data_offset": 2048, 00:11:00.922 "data_size": 63488 00:11:00.922 }, 00:11:00.922 { 00:11:00.922 "name": "pt4", 00:11:00.923 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:00.923 "is_configured": true, 00:11:00.923 "data_offset": 2048, 00:11:00.923 "data_size": 63488 00:11:00.923 } 00:11:00.923 ] 00:11:00.923 }' 00:11:00.923 15:15:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:00.923 15:15:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.182 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:01.182 15:15:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.182 15:15:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.182 [2024-11-27 15:15:29.228872] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:01.182 [2024-11-27 15:15:29.228919] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:01.182 [2024-11-27 15:15:29.229007] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:01.182 [2024-11-27 15:15:29.229079] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:01.182 [2024-11-27 15:15:29.229094] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:11:01.183 15:15:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.183 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:11:01.183 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:01.183 15:15:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.183 15:15:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.183 15:15:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.183 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:11:01.183 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:11:01.183 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:11:01.183 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:11:01.183 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:11:01.183 15:15:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.183 15:15:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.183 15:15:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.183 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:11:01.183 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:11:01.183 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:11:01.183 15:15:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.183 15:15:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.442 15:15:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.442 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:11:01.442 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:11:01.442 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt4 00:11:01.442 15:15:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.442 15:15:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.442 15:15:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.442 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:11:01.442 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:11:01.442 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:11:01.442 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:11:01.442 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:11:01.442 15:15:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.442 15:15:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.442 [2024-11-27 15:15:29.312695] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:11:01.442 [2024-11-27 15:15:29.312761] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:01.442 [2024-11-27 15:15:29.312778] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:11:01.442 [2024-11-27 15:15:29.312790] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:01.442 [2024-11-27 15:15:29.315049] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:01.442 [2024-11-27 15:15:29.315090] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:11:01.442 [2024-11-27 15:15:29.315177] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:11:01.442 [2024-11-27 15:15:29.315213] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:11:01.442 pt2 00:11:01.443 15:15:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.443 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:11:01.443 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:01.443 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:01.443 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:01.443 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:01.443 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:01.443 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:01.443 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:01.443 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:01.443 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:01.443 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:01.443 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:01.443 15:15:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.443 15:15:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.443 15:15:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.443 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:01.443 "name": "raid_bdev1", 00:11:01.443 "uuid": "04fb4881-7d28-4de8-97e1-80b94e75d55b", 00:11:01.443 "strip_size_kb": 0, 00:11:01.443 "state": "configuring", 00:11:01.443 "raid_level": "raid1", 00:11:01.443 "superblock": true, 00:11:01.443 "num_base_bdevs": 4, 00:11:01.443 "num_base_bdevs_discovered": 1, 00:11:01.443 "num_base_bdevs_operational": 3, 00:11:01.443 "base_bdevs_list": [ 00:11:01.443 { 00:11:01.443 "name": null, 00:11:01.443 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:01.443 "is_configured": false, 00:11:01.443 "data_offset": 2048, 00:11:01.443 "data_size": 63488 00:11:01.443 }, 00:11:01.443 { 00:11:01.443 "name": "pt2", 00:11:01.443 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:01.443 "is_configured": true, 00:11:01.443 "data_offset": 2048, 00:11:01.443 "data_size": 63488 00:11:01.443 }, 00:11:01.443 { 00:11:01.443 "name": null, 00:11:01.443 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:01.443 "is_configured": false, 00:11:01.443 "data_offset": 2048, 00:11:01.443 "data_size": 63488 00:11:01.443 }, 00:11:01.443 { 00:11:01.443 "name": null, 00:11:01.443 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:01.443 "is_configured": false, 00:11:01.443 "data_offset": 2048, 00:11:01.443 "data_size": 63488 00:11:01.443 } 00:11:01.443 ] 00:11:01.443 }' 00:11:01.443 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:01.443 15:15:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.702 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:11:01.702 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:11:01.702 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:11:01.702 15:15:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.702 15:15:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.702 [2024-11-27 15:15:29.740010] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:11:01.702 [2024-11-27 15:15:29.740086] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:01.702 [2024-11-27 15:15:29.740106] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:11:01.702 [2024-11-27 15:15:29.740121] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:01.702 [2024-11-27 15:15:29.740544] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:01.702 [2024-11-27 15:15:29.740575] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:11:01.702 [2024-11-27 15:15:29.740652] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:11:01.702 [2024-11-27 15:15:29.740683] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:11:01.702 pt3 00:11:01.702 15:15:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.702 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:11:01.702 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:01.702 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:01.702 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:01.702 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:01.702 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:01.702 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:01.702 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:01.702 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:01.702 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:01.702 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:01.702 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:01.702 15:15:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.702 15:15:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.702 15:15:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.702 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:01.702 "name": "raid_bdev1", 00:11:01.702 "uuid": "04fb4881-7d28-4de8-97e1-80b94e75d55b", 00:11:01.702 "strip_size_kb": 0, 00:11:01.702 "state": "configuring", 00:11:01.702 "raid_level": "raid1", 00:11:01.702 "superblock": true, 00:11:01.702 "num_base_bdevs": 4, 00:11:01.702 "num_base_bdevs_discovered": 2, 00:11:01.702 "num_base_bdevs_operational": 3, 00:11:01.702 "base_bdevs_list": [ 00:11:01.702 { 00:11:01.702 "name": null, 00:11:01.702 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:01.702 "is_configured": false, 00:11:01.702 "data_offset": 2048, 00:11:01.702 "data_size": 63488 00:11:01.702 }, 00:11:01.702 { 00:11:01.702 "name": "pt2", 00:11:01.702 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:01.702 "is_configured": true, 00:11:01.702 "data_offset": 2048, 00:11:01.702 "data_size": 63488 00:11:01.702 }, 00:11:01.702 { 00:11:01.702 "name": "pt3", 00:11:01.702 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:01.702 "is_configured": true, 00:11:01.702 "data_offset": 2048, 00:11:01.702 "data_size": 63488 00:11:01.702 }, 00:11:01.702 { 00:11:01.702 "name": null, 00:11:01.702 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:01.702 "is_configured": false, 00:11:01.702 "data_offset": 2048, 00:11:01.702 "data_size": 63488 00:11:01.702 } 00:11:01.702 ] 00:11:01.702 }' 00:11:01.702 15:15:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:01.702 15:15:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.271 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:11:02.271 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:11:02.271 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=3 00:11:02.271 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:11:02.271 15:15:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:02.271 15:15:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.271 [2024-11-27 15:15:30.167270] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:11:02.271 [2024-11-27 15:15:30.167456] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:02.271 [2024-11-27 15:15:30.167483] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:11:02.271 [2024-11-27 15:15:30.167495] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:02.271 [2024-11-27 15:15:30.167939] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:02.271 [2024-11-27 15:15:30.167961] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:11:02.271 [2024-11-27 15:15:30.168044] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:11:02.271 [2024-11-27 15:15:30.168071] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:11:02.271 [2024-11-27 15:15:30.168173] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:11:02.271 [2024-11-27 15:15:30.168185] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:02.271 [2024-11-27 15:15:30.168425] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:11:02.271 [2024-11-27 15:15:30.168560] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:11:02.271 [2024-11-27 15:15:30.168569] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006d00 00:11:02.271 [2024-11-27 15:15:30.168677] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:02.271 pt4 00:11:02.271 15:15:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.271 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:11:02.271 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:02.271 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:02.271 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:02.271 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:02.271 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:02.271 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:02.271 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:02.271 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:02.271 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:02.271 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:02.271 15:15:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:02.271 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:02.271 15:15:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.271 15:15:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.271 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:02.271 "name": "raid_bdev1", 00:11:02.271 "uuid": "04fb4881-7d28-4de8-97e1-80b94e75d55b", 00:11:02.271 "strip_size_kb": 0, 00:11:02.271 "state": "online", 00:11:02.271 "raid_level": "raid1", 00:11:02.271 "superblock": true, 00:11:02.271 "num_base_bdevs": 4, 00:11:02.271 "num_base_bdevs_discovered": 3, 00:11:02.271 "num_base_bdevs_operational": 3, 00:11:02.271 "base_bdevs_list": [ 00:11:02.271 { 00:11:02.271 "name": null, 00:11:02.271 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:02.271 "is_configured": false, 00:11:02.271 "data_offset": 2048, 00:11:02.271 "data_size": 63488 00:11:02.271 }, 00:11:02.271 { 00:11:02.271 "name": "pt2", 00:11:02.271 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:02.271 "is_configured": true, 00:11:02.271 "data_offset": 2048, 00:11:02.271 "data_size": 63488 00:11:02.271 }, 00:11:02.271 { 00:11:02.271 "name": "pt3", 00:11:02.271 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:02.271 "is_configured": true, 00:11:02.271 "data_offset": 2048, 00:11:02.271 "data_size": 63488 00:11:02.271 }, 00:11:02.271 { 00:11:02.271 "name": "pt4", 00:11:02.271 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:02.271 "is_configured": true, 00:11:02.271 "data_offset": 2048, 00:11:02.271 "data_size": 63488 00:11:02.271 } 00:11:02.271 ] 00:11:02.271 }' 00:11:02.271 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:02.271 15:15:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.531 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:02.531 15:15:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:02.531 15:15:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.531 [2024-11-27 15:15:30.614556] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:02.531 [2024-11-27 15:15:30.614683] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:02.531 [2024-11-27 15:15:30.614776] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:02.531 [2024-11-27 15:15:30.614869] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:02.531 [2024-11-27 15:15:30.614970] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name raid_bdev1, state offline 00:11:02.531 15:15:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.531 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:02.531 15:15:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:02.531 15:15:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.531 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:11:02.531 15:15:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.790 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:11:02.790 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:11:02.790 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 4 -gt 2 ']' 00:11:02.790 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@534 -- # i=3 00:11:02.790 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt4 00:11:02.790 15:15:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:02.790 15:15:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.790 15:15:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.790 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:11:02.790 15:15:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:02.790 15:15:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.790 [2024-11-27 15:15:30.690408] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:11:02.790 [2024-11-27 15:15:30.690540] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:02.790 [2024-11-27 15:15:30.690577] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c080 00:11:02.790 [2024-11-27 15:15:30.690604] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:02.790 [2024-11-27 15:15:30.692799] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:02.790 [2024-11-27 15:15:30.692874] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:11:02.790 [2024-11-27 15:15:30.692996] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:11:02.790 [2024-11-27 15:15:30.693069] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:11:02.790 [2024-11-27 15:15:30.693213] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:11:02.790 [2024-11-27 15:15:30.693267] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:02.790 [2024-11-27 15:15:30.693317] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007080 name raid_bdev1, state configuring 00:11:02.790 [2024-11-27 15:15:30.693392] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:11:02.790 [2024-11-27 15:15:30.693526] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:11:02.790 pt1 00:11:02.790 15:15:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.790 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 4 -gt 2 ']' 00:11:02.790 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:11:02.790 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:02.790 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:02.790 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:02.790 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:02.790 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:02.790 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:02.790 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:02.790 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:02.790 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:02.790 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:02.790 15:15:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:02.790 15:15:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.790 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:02.790 15:15:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.790 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:02.790 "name": "raid_bdev1", 00:11:02.790 "uuid": "04fb4881-7d28-4de8-97e1-80b94e75d55b", 00:11:02.790 "strip_size_kb": 0, 00:11:02.790 "state": "configuring", 00:11:02.790 "raid_level": "raid1", 00:11:02.790 "superblock": true, 00:11:02.790 "num_base_bdevs": 4, 00:11:02.790 "num_base_bdevs_discovered": 2, 00:11:02.790 "num_base_bdevs_operational": 3, 00:11:02.790 "base_bdevs_list": [ 00:11:02.790 { 00:11:02.790 "name": null, 00:11:02.790 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:02.790 "is_configured": false, 00:11:02.790 "data_offset": 2048, 00:11:02.790 "data_size": 63488 00:11:02.790 }, 00:11:02.790 { 00:11:02.790 "name": "pt2", 00:11:02.790 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:02.790 "is_configured": true, 00:11:02.790 "data_offset": 2048, 00:11:02.790 "data_size": 63488 00:11:02.790 }, 00:11:02.790 { 00:11:02.790 "name": "pt3", 00:11:02.790 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:02.790 "is_configured": true, 00:11:02.790 "data_offset": 2048, 00:11:02.790 "data_size": 63488 00:11:02.790 }, 00:11:02.790 { 00:11:02.790 "name": null, 00:11:02.790 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:02.790 "is_configured": false, 00:11:02.790 "data_offset": 2048, 00:11:02.790 "data_size": 63488 00:11:02.790 } 00:11:02.790 ] 00:11:02.790 }' 00:11:02.790 15:15:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:02.790 15:15:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.050 15:15:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:11:03.050 15:15:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:11:03.050 15:15:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.050 15:15:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.050 15:15:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.309 15:15:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:11:03.309 15:15:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:11:03.309 15:15:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.309 15:15:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.309 [2024-11-27 15:15:31.181587] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:11:03.309 [2024-11-27 15:15:31.181656] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:03.309 [2024-11-27 15:15:31.181677] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c680 00:11:03.309 [2024-11-27 15:15:31.181688] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:03.309 [2024-11-27 15:15:31.182105] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:03.309 [2024-11-27 15:15:31.182125] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:11:03.309 [2024-11-27 15:15:31.182195] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:11:03.309 [2024-11-27 15:15:31.182221] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:11:03.309 [2024-11-27 15:15:31.182319] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007400 00:11:03.309 [2024-11-27 15:15:31.182339] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:03.309 [2024-11-27 15:15:31.182567] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:11:03.309 [2024-11-27 15:15:31.182685] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007400 00:11:03.309 [2024-11-27 15:15:31.182693] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007400 00:11:03.309 [2024-11-27 15:15:31.182798] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:03.309 pt4 00:11:03.309 15:15:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.309 15:15:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:11:03.309 15:15:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:03.309 15:15:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:03.309 15:15:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:03.309 15:15:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:03.309 15:15:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:03.309 15:15:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:03.309 15:15:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:03.309 15:15:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:03.309 15:15:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:03.309 15:15:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:03.309 15:15:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.309 15:15:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.309 15:15:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:03.309 15:15:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.309 15:15:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:03.309 "name": "raid_bdev1", 00:11:03.309 "uuid": "04fb4881-7d28-4de8-97e1-80b94e75d55b", 00:11:03.309 "strip_size_kb": 0, 00:11:03.309 "state": "online", 00:11:03.309 "raid_level": "raid1", 00:11:03.309 "superblock": true, 00:11:03.309 "num_base_bdevs": 4, 00:11:03.309 "num_base_bdevs_discovered": 3, 00:11:03.309 "num_base_bdevs_operational": 3, 00:11:03.309 "base_bdevs_list": [ 00:11:03.309 { 00:11:03.309 "name": null, 00:11:03.309 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:03.309 "is_configured": false, 00:11:03.309 "data_offset": 2048, 00:11:03.309 "data_size": 63488 00:11:03.309 }, 00:11:03.309 { 00:11:03.309 "name": "pt2", 00:11:03.309 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:03.309 "is_configured": true, 00:11:03.309 "data_offset": 2048, 00:11:03.309 "data_size": 63488 00:11:03.309 }, 00:11:03.309 { 00:11:03.309 "name": "pt3", 00:11:03.309 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:03.309 "is_configured": true, 00:11:03.309 "data_offset": 2048, 00:11:03.309 "data_size": 63488 00:11:03.309 }, 00:11:03.309 { 00:11:03.309 "name": "pt4", 00:11:03.309 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:03.309 "is_configured": true, 00:11:03.309 "data_offset": 2048, 00:11:03.309 "data_size": 63488 00:11:03.309 } 00:11:03.309 ] 00:11:03.309 }' 00:11:03.309 15:15:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:03.309 15:15:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.568 15:15:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:11:03.568 15:15:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.568 15:15:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.568 15:15:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:11:03.568 15:15:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.568 15:15:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:11:03.568 15:15:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:03.568 15:15:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.568 15:15:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.568 15:15:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:11:03.568 [2024-11-27 15:15:31.633181] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:03.568 15:15:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.568 15:15:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 04fb4881-7d28-4de8-97e1-80b94e75d55b '!=' 04fb4881-7d28-4de8-97e1-80b94e75d55b ']' 00:11:03.568 15:15:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 85420 00:11:03.568 15:15:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 85420 ']' 00:11:03.568 15:15:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 85420 00:11:03.568 15:15:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:11:03.568 15:15:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:03.827 15:15:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 85420 00:11:03.827 15:15:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:03.827 15:15:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:03.827 15:15:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 85420' 00:11:03.827 killing process with pid 85420 00:11:03.827 15:15:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 85420 00:11:03.827 [2024-11-27 15:15:31.704329] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:03.827 [2024-11-27 15:15:31.704487] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:03.827 15:15:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 85420 00:11:03.828 [2024-11-27 15:15:31.704572] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:03.828 [2024-11-27 15:15:31.704584] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007400 name raid_bdev1, state offline 00:11:03.828 [2024-11-27 15:15:31.747966] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:04.087 ************************************ 00:11:04.087 END TEST raid_superblock_test 00:11:04.087 ************************************ 00:11:04.087 15:15:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:11:04.087 00:11:04.087 real 0m6.888s 00:11:04.087 user 0m11.574s 00:11:04.087 sys 0m1.517s 00:11:04.087 15:15:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:04.088 15:15:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.088 15:15:32 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 4 read 00:11:04.088 15:15:32 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:11:04.088 15:15:32 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:04.088 15:15:32 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:04.088 ************************************ 00:11:04.088 START TEST raid_read_error_test 00:11:04.088 ************************************ 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 4 read 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.Uu8BRvxwAD 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=85896 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 85896 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 85896 ']' 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:04.088 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:04.088 15:15:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.088 [2024-11-27 15:15:32.139515] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:11:04.088 [2024-11-27 15:15:32.139740] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid85896 ] 00:11:04.348 [2024-11-27 15:15:32.288276] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:04.348 [2024-11-27 15:15:32.313450] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:04.348 [2024-11-27 15:15:32.356153] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:04.348 [2024-11-27 15:15:32.356185] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:04.916 15:15:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:04.916 15:15:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:11:04.916 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:04.916 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:04.916 15:15:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:04.916 15:15:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.916 BaseBdev1_malloc 00:11:04.916 15:15:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:04.916 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:11:04.916 15:15:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:04.916 15:15:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.916 true 00:11:04.916 15:15:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:04.916 15:15:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:11:04.916 15:15:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:04.916 15:15:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.916 [2024-11-27 15:15:33.004089] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:11:04.916 [2024-11-27 15:15:33.004240] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:04.916 [2024-11-27 15:15:33.004287] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:11:04.916 [2024-11-27 15:15:33.004327] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:04.916 [2024-11-27 15:15:33.006533] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:04.916 [2024-11-27 15:15:33.006605] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:04.916 BaseBdev1 00:11:04.916 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:04.916 15:15:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:04.916 15:15:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:04.916 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:04.916 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.176 BaseBdev2_malloc 00:11:05.176 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.176 15:15:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:11:05.176 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.176 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.176 true 00:11:05.176 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.176 15:15:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:11:05.176 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.176 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.176 [2024-11-27 15:15:33.045007] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:11:05.176 [2024-11-27 15:15:33.045144] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:05.176 [2024-11-27 15:15:33.045179] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:11:05.176 [2024-11-27 15:15:33.045215] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:05.176 [2024-11-27 15:15:33.047179] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:05.176 [2024-11-27 15:15:33.047252] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:05.176 BaseBdev2 00:11:05.176 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.176 15:15:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:05.176 15:15:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:11:05.176 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.176 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.176 BaseBdev3_malloc 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.177 true 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.177 [2024-11-27 15:15:33.085667] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:11:05.177 [2024-11-27 15:15:33.085726] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:05.177 [2024-11-27 15:15:33.085747] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:11:05.177 [2024-11-27 15:15:33.085755] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:05.177 [2024-11-27 15:15:33.087817] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:05.177 [2024-11-27 15:15:33.087855] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:11:05.177 BaseBdev3 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.177 BaseBdev4_malloc 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.177 true 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.177 [2024-11-27 15:15:33.136019] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:11:05.177 [2024-11-27 15:15:33.136079] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:05.177 [2024-11-27 15:15:33.136117] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:11:05.177 [2024-11-27 15:15:33.136126] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:05.177 [2024-11-27 15:15:33.138077] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:05.177 [2024-11-27 15:15:33.138190] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:11:05.177 BaseBdev4 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.177 [2024-11-27 15:15:33.148056] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:05.177 [2024-11-27 15:15:33.149785] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:05.177 [2024-11-27 15:15:33.149873] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:05.177 [2024-11-27 15:15:33.149934] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:05.177 [2024-11-27 15:15:33.150127] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007080 00:11:05.177 [2024-11-27 15:15:33.150147] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:05.177 [2024-11-27 15:15:33.150384] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006700 00:11:05.177 [2024-11-27 15:15:33.150530] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007080 00:11:05.177 [2024-11-27 15:15:33.150549] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007080 00:11:05.177 [2024-11-27 15:15:33.150672] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:05.177 "name": "raid_bdev1", 00:11:05.177 "uuid": "d52c8d8f-e640-475c-a006-dfc2dd52f988", 00:11:05.177 "strip_size_kb": 0, 00:11:05.177 "state": "online", 00:11:05.177 "raid_level": "raid1", 00:11:05.177 "superblock": true, 00:11:05.177 "num_base_bdevs": 4, 00:11:05.177 "num_base_bdevs_discovered": 4, 00:11:05.177 "num_base_bdevs_operational": 4, 00:11:05.177 "base_bdevs_list": [ 00:11:05.177 { 00:11:05.177 "name": "BaseBdev1", 00:11:05.177 "uuid": "32eeeae1-1d4d-504d-9b5e-2eecc4899d9f", 00:11:05.177 "is_configured": true, 00:11:05.177 "data_offset": 2048, 00:11:05.177 "data_size": 63488 00:11:05.177 }, 00:11:05.177 { 00:11:05.177 "name": "BaseBdev2", 00:11:05.177 "uuid": "e86c1639-d1b7-55e2-a843-1c1d7b4092d2", 00:11:05.177 "is_configured": true, 00:11:05.177 "data_offset": 2048, 00:11:05.177 "data_size": 63488 00:11:05.177 }, 00:11:05.177 { 00:11:05.177 "name": "BaseBdev3", 00:11:05.177 "uuid": "565ce4b4-c049-5b28-b5cb-8ef326faa2de", 00:11:05.177 "is_configured": true, 00:11:05.177 "data_offset": 2048, 00:11:05.177 "data_size": 63488 00:11:05.177 }, 00:11:05.177 { 00:11:05.177 "name": "BaseBdev4", 00:11:05.177 "uuid": "da0ebed3-0e98-5499-b4ca-49b7a50d9ef2", 00:11:05.177 "is_configured": true, 00:11:05.177 "data_offset": 2048, 00:11:05.177 "data_size": 63488 00:11:05.177 } 00:11:05.177 ] 00:11:05.177 }' 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:05.177 15:15:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.746 15:15:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:11:05.746 15:15:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:11:05.746 [2024-11-27 15:15:33.707517] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000068a0 00:11:06.684 15:15:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:11:06.684 15:15:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.684 15:15:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.684 15:15:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.684 15:15:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:11:06.684 15:15:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:11:06.685 15:15:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:11:06.685 15:15:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:11:06.685 15:15:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:11:06.685 15:15:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:06.685 15:15:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:06.685 15:15:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:06.685 15:15:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:06.685 15:15:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:06.685 15:15:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:06.685 15:15:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:06.685 15:15:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:06.685 15:15:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:06.685 15:15:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:06.685 15:15:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:06.685 15:15:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.685 15:15:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.685 15:15:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.685 15:15:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:06.685 "name": "raid_bdev1", 00:11:06.685 "uuid": "d52c8d8f-e640-475c-a006-dfc2dd52f988", 00:11:06.685 "strip_size_kb": 0, 00:11:06.685 "state": "online", 00:11:06.685 "raid_level": "raid1", 00:11:06.685 "superblock": true, 00:11:06.685 "num_base_bdevs": 4, 00:11:06.685 "num_base_bdevs_discovered": 4, 00:11:06.685 "num_base_bdevs_operational": 4, 00:11:06.685 "base_bdevs_list": [ 00:11:06.685 { 00:11:06.685 "name": "BaseBdev1", 00:11:06.685 "uuid": "32eeeae1-1d4d-504d-9b5e-2eecc4899d9f", 00:11:06.685 "is_configured": true, 00:11:06.685 "data_offset": 2048, 00:11:06.685 "data_size": 63488 00:11:06.685 }, 00:11:06.685 { 00:11:06.685 "name": "BaseBdev2", 00:11:06.685 "uuid": "e86c1639-d1b7-55e2-a843-1c1d7b4092d2", 00:11:06.685 "is_configured": true, 00:11:06.685 "data_offset": 2048, 00:11:06.685 "data_size": 63488 00:11:06.685 }, 00:11:06.685 { 00:11:06.685 "name": "BaseBdev3", 00:11:06.685 "uuid": "565ce4b4-c049-5b28-b5cb-8ef326faa2de", 00:11:06.685 "is_configured": true, 00:11:06.685 "data_offset": 2048, 00:11:06.685 "data_size": 63488 00:11:06.685 }, 00:11:06.685 { 00:11:06.685 "name": "BaseBdev4", 00:11:06.685 "uuid": "da0ebed3-0e98-5499-b4ca-49b7a50d9ef2", 00:11:06.685 "is_configured": true, 00:11:06.685 "data_offset": 2048, 00:11:06.685 "data_size": 63488 00:11:06.685 } 00:11:06.685 ] 00:11:06.685 }' 00:11:06.685 15:15:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:06.685 15:15:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.254 15:15:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:07.254 15:15:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:07.254 15:15:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.254 [2024-11-27 15:15:35.058216] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:07.254 [2024-11-27 15:15:35.058361] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:07.254 [2024-11-27 15:15:35.061081] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:07.254 [2024-11-27 15:15:35.061204] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:07.254 [2024-11-27 15:15:35.061354] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:07.254 [2024-11-27 15:15:35.061401] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007080 name raid_bdev1, state offline 00:11:07.254 { 00:11:07.254 "results": [ 00:11:07.254 { 00:11:07.254 "job": "raid_bdev1", 00:11:07.254 "core_mask": "0x1", 00:11:07.254 "workload": "randrw", 00:11:07.254 "percentage": 50, 00:11:07.254 "status": "finished", 00:11:07.254 "queue_depth": 1, 00:11:07.254 "io_size": 131072, 00:11:07.254 "runtime": 1.351814, 00:11:07.254 "iops": 11371.38689198366, 00:11:07.254 "mibps": 1421.4233614979576, 00:11:07.254 "io_failed": 0, 00:11:07.254 "io_timeout": 0, 00:11:07.254 "avg_latency_us": 85.37194581653026, 00:11:07.254 "min_latency_us": 22.805240174672488, 00:11:07.254 "max_latency_us": 1459.5353711790392 00:11:07.254 } 00:11:07.254 ], 00:11:07.254 "core_count": 1 00:11:07.254 } 00:11:07.254 15:15:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:07.254 15:15:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 85896 00:11:07.254 15:15:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 85896 ']' 00:11:07.254 15:15:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 85896 00:11:07.254 15:15:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:11:07.254 15:15:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:07.254 15:15:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 85896 00:11:07.254 killing process with pid 85896 00:11:07.254 15:15:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:07.254 15:15:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:07.254 15:15:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 85896' 00:11:07.254 15:15:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 85896 00:11:07.254 [2024-11-27 15:15:35.095101] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:07.254 15:15:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 85896 00:11:07.254 [2024-11-27 15:15:35.130299] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:07.254 15:15:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.Uu8BRvxwAD 00:11:07.254 15:15:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:11:07.254 15:15:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:11:07.254 15:15:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:11:07.254 15:15:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:11:07.254 15:15:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:07.254 15:15:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:11:07.254 ************************************ 00:11:07.254 END TEST raid_read_error_test 00:11:07.254 ************************************ 00:11:07.254 15:15:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:11:07.254 00:11:07.254 real 0m3.318s 00:11:07.254 user 0m4.179s 00:11:07.254 sys 0m0.546s 00:11:07.254 15:15:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:07.254 15:15:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.514 15:15:35 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 4 write 00:11:07.514 15:15:35 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:11:07.514 15:15:35 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:07.514 15:15:35 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:07.514 ************************************ 00:11:07.514 START TEST raid_write_error_test 00:11:07.514 ************************************ 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 4 write 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.01X4UiMe6n 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=86030 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 86030 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 86030 ']' 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:07.514 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:07.514 15:15:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.514 [2024-11-27 15:15:35.544256] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:11:07.514 [2024-11-27 15:15:35.544512] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid86030 ] 00:11:07.774 [2024-11-27 15:15:35.720140] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:07.774 [2024-11-27 15:15:35.748145] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:07.774 [2024-11-27 15:15:35.792895] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:07.774 [2024-11-27 15:15:35.793039] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:08.342 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:08.342 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:11:08.342 15:15:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:08.342 15:15:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:08.342 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.342 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.342 BaseBdev1_malloc 00:11:08.342 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.342 15:15:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:11:08.342 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.342 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.342 true 00:11:08.342 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.342 15:15:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:11:08.342 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.342 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.342 [2024-11-27 15:15:36.434365] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:11:08.342 [2024-11-27 15:15:36.434530] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:08.342 [2024-11-27 15:15:36.434587] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:11:08.342 [2024-11-27 15:15:36.434621] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:08.342 [2024-11-27 15:15:36.437060] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:08.342 [2024-11-27 15:15:36.437143] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:08.342 BaseBdev1 00:11:08.342 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.342 15:15:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:08.342 15:15:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:08.342 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.342 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.602 BaseBdev2_malloc 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.602 true 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.602 [2024-11-27 15:15:36.475616] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:11:08.602 [2024-11-27 15:15:36.475754] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:08.602 [2024-11-27 15:15:36.475779] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:11:08.602 [2024-11-27 15:15:36.475790] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:08.602 [2024-11-27 15:15:36.478169] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:08.602 [2024-11-27 15:15:36.478208] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:08.602 BaseBdev2 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.602 BaseBdev3_malloc 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.602 true 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.602 [2024-11-27 15:15:36.516769] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:11:08.602 [2024-11-27 15:15:36.516832] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:08.602 [2024-11-27 15:15:36.516853] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:11:08.602 [2024-11-27 15:15:36.516864] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:08.602 [2024-11-27 15:15:36.519176] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:08.602 [2024-11-27 15:15:36.519217] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:11:08.602 BaseBdev3 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.602 BaseBdev4_malloc 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.602 true 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.602 [2024-11-27 15:15:36.566404] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:11:08.602 [2024-11-27 15:15:36.566468] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:08.602 [2024-11-27 15:15:36.566492] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:11:08.602 [2024-11-27 15:15:36.566502] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:08.602 [2024-11-27 15:15:36.568774] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:08.602 [2024-11-27 15:15:36.568872] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:11:08.602 BaseBdev4 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.602 [2024-11-27 15:15:36.578450] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:08.602 [2024-11-27 15:15:36.580545] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:08.602 [2024-11-27 15:15:36.580646] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:08.602 [2024-11-27 15:15:36.580707] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:08.602 [2024-11-27 15:15:36.580949] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007080 00:11:08.602 [2024-11-27 15:15:36.580965] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:08.602 [2024-11-27 15:15:36.581234] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006700 00:11:08.602 [2024-11-27 15:15:36.581400] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007080 00:11:08.602 [2024-11-27 15:15:36.581423] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007080 00:11:08.602 [2024-11-27 15:15:36.581559] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:08.602 15:15:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:08.603 15:15:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:08.603 15:15:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:08.603 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.603 15:15:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:08.603 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.603 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.603 15:15:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:08.603 "name": "raid_bdev1", 00:11:08.603 "uuid": "20757c0e-3b38-4607-9c82-c4222326dbf5", 00:11:08.603 "strip_size_kb": 0, 00:11:08.603 "state": "online", 00:11:08.603 "raid_level": "raid1", 00:11:08.603 "superblock": true, 00:11:08.603 "num_base_bdevs": 4, 00:11:08.603 "num_base_bdevs_discovered": 4, 00:11:08.603 "num_base_bdevs_operational": 4, 00:11:08.603 "base_bdevs_list": [ 00:11:08.603 { 00:11:08.603 "name": "BaseBdev1", 00:11:08.603 "uuid": "96df0652-a10d-549e-af1a-4a8147b24a90", 00:11:08.603 "is_configured": true, 00:11:08.603 "data_offset": 2048, 00:11:08.603 "data_size": 63488 00:11:08.603 }, 00:11:08.603 { 00:11:08.603 "name": "BaseBdev2", 00:11:08.603 "uuid": "e747d11f-bc66-5841-9311-b76d010b5a35", 00:11:08.603 "is_configured": true, 00:11:08.603 "data_offset": 2048, 00:11:08.603 "data_size": 63488 00:11:08.603 }, 00:11:08.603 { 00:11:08.603 "name": "BaseBdev3", 00:11:08.603 "uuid": "49681900-8d2c-57cd-a28e-0f36fbd3f6e3", 00:11:08.603 "is_configured": true, 00:11:08.603 "data_offset": 2048, 00:11:08.603 "data_size": 63488 00:11:08.603 }, 00:11:08.603 { 00:11:08.603 "name": "BaseBdev4", 00:11:08.603 "uuid": "3ba70109-7ae2-5a7c-9d30-f97a0b3a61ee", 00:11:08.603 "is_configured": true, 00:11:08.603 "data_offset": 2048, 00:11:08.603 "data_size": 63488 00:11:08.603 } 00:11:08.603 ] 00:11:08.603 }' 00:11:08.603 15:15:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:08.603 15:15:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.171 15:15:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:11:09.171 15:15:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:11:09.171 [2024-11-27 15:15:37.130044] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000068a0 00:11:10.109 15:15:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:11:10.109 15:15:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:10.109 15:15:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:10.109 [2024-11-27 15:15:38.045500] bdev_raid.c:2276:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:11:10.109 [2024-11-27 15:15:38.045585] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:10.109 [2024-11-27 15:15:38.045820] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d0000068a0 00:11:10.109 15:15:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:10.109 15:15:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:11:10.109 15:15:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:11:10.109 15:15:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:11:10.109 15:15:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=3 00:11:10.109 15:15:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:11:10.109 15:15:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:10.109 15:15:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:10.109 15:15:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:10.109 15:15:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:10.109 15:15:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:10.109 15:15:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:10.109 15:15:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:10.109 15:15:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:10.109 15:15:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:10.109 15:15:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:10.109 15:15:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:10.109 15:15:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:10.109 15:15:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:10.109 15:15:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:10.109 15:15:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:10.109 "name": "raid_bdev1", 00:11:10.109 "uuid": "20757c0e-3b38-4607-9c82-c4222326dbf5", 00:11:10.109 "strip_size_kb": 0, 00:11:10.109 "state": "online", 00:11:10.109 "raid_level": "raid1", 00:11:10.109 "superblock": true, 00:11:10.109 "num_base_bdevs": 4, 00:11:10.109 "num_base_bdevs_discovered": 3, 00:11:10.109 "num_base_bdevs_operational": 3, 00:11:10.109 "base_bdevs_list": [ 00:11:10.109 { 00:11:10.109 "name": null, 00:11:10.109 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:10.109 "is_configured": false, 00:11:10.109 "data_offset": 0, 00:11:10.109 "data_size": 63488 00:11:10.109 }, 00:11:10.109 { 00:11:10.109 "name": "BaseBdev2", 00:11:10.109 "uuid": "e747d11f-bc66-5841-9311-b76d010b5a35", 00:11:10.109 "is_configured": true, 00:11:10.109 "data_offset": 2048, 00:11:10.109 "data_size": 63488 00:11:10.109 }, 00:11:10.109 { 00:11:10.109 "name": "BaseBdev3", 00:11:10.109 "uuid": "49681900-8d2c-57cd-a28e-0f36fbd3f6e3", 00:11:10.109 "is_configured": true, 00:11:10.109 "data_offset": 2048, 00:11:10.109 "data_size": 63488 00:11:10.109 }, 00:11:10.109 { 00:11:10.109 "name": "BaseBdev4", 00:11:10.109 "uuid": "3ba70109-7ae2-5a7c-9d30-f97a0b3a61ee", 00:11:10.109 "is_configured": true, 00:11:10.109 "data_offset": 2048, 00:11:10.109 "data_size": 63488 00:11:10.109 } 00:11:10.109 ] 00:11:10.109 }' 00:11:10.109 15:15:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:10.109 15:15:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:10.678 15:15:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:10.678 15:15:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:10.678 15:15:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:10.678 [2024-11-27 15:15:38.541687] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:10.678 [2024-11-27 15:15:38.541828] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:10.678 [2024-11-27 15:15:38.544857] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:10.678 [2024-11-27 15:15:38.544988] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:10.678 [2024-11-27 15:15:38.545137] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:10.678 [2024-11-27 15:15:38.545194] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007080 name raid_bdev1, state offline 00:11:10.678 { 00:11:10.678 "results": [ 00:11:10.678 { 00:11:10.678 "job": "raid_bdev1", 00:11:10.678 "core_mask": "0x1", 00:11:10.678 "workload": "randrw", 00:11:10.678 "percentage": 50, 00:11:10.678 "status": "finished", 00:11:10.678 "queue_depth": 1, 00:11:10.678 "io_size": 131072, 00:11:10.678 "runtime": 1.412387, 00:11:10.678 "iops": 11093.276842678388, 00:11:10.678 "mibps": 1386.6596053347985, 00:11:10.678 "io_failed": 0, 00:11:10.678 "io_timeout": 0, 00:11:10.678 "avg_latency_us": 87.05358637135407, 00:11:10.678 "min_latency_us": 24.929257641921396, 00:11:10.678 "max_latency_us": 1681.3275109170306 00:11:10.678 } 00:11:10.678 ], 00:11:10.678 "core_count": 1 00:11:10.678 } 00:11:10.678 15:15:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:10.678 15:15:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 86030 00:11:10.678 15:15:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 86030 ']' 00:11:10.678 15:15:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 86030 00:11:10.678 15:15:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:11:10.678 15:15:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:10.678 15:15:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 86030 00:11:10.678 killing process with pid 86030 00:11:10.678 15:15:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:10.678 15:15:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:10.678 15:15:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 86030' 00:11:10.678 15:15:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 86030 00:11:10.678 [2024-11-27 15:15:38.591249] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:10.678 15:15:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 86030 00:11:10.678 [2024-11-27 15:15:38.628568] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:10.938 15:15:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:11:10.938 15:15:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.01X4UiMe6n 00:11:10.938 15:15:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:11:10.938 ************************************ 00:11:10.938 END TEST raid_write_error_test 00:11:10.938 ************************************ 00:11:10.938 15:15:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:11:10.938 15:15:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:11:10.938 15:15:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:10.938 15:15:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:11:10.938 15:15:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:11:10.938 00:11:10.938 real 0m3.426s 00:11:10.938 user 0m4.339s 00:11:10.938 sys 0m0.606s 00:11:10.938 15:15:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:10.938 15:15:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:10.938 15:15:38 bdev_raid -- bdev/bdev_raid.sh@976 -- # '[' true = true ']' 00:11:10.938 15:15:38 bdev_raid -- bdev/bdev_raid.sh@977 -- # for n in 2 4 00:11:10.939 15:15:38 bdev_raid -- bdev/bdev_raid.sh@978 -- # run_test raid_rebuild_test raid_rebuild_test raid1 2 false false true 00:11:10.939 15:15:38 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:11:10.939 15:15:38 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:10.939 15:15:38 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:10.939 ************************************ 00:11:10.939 START TEST raid_rebuild_test 00:11:10.939 ************************************ 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 false false true 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=86158 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 86158 00:11:10.939 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@835 -- # '[' -z 86158 ']' 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:10.939 15:15:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:11.205 I/O size of 3145728 is greater than zero copy threshold (65536). 00:11:11.205 Zero copy mechanism will not be used. 00:11:11.205 [2024-11-27 15:15:39.050051] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:11:11.206 [2024-11-27 15:15:39.050215] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid86158 ] 00:11:11.206 [2024-11-27 15:15:39.229523] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:11.206 [2024-11-27 15:15:39.257488] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:11.206 [2024-11-27 15:15:39.301433] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:11.206 [2024-11-27 15:15:39.301474] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:12.146 15:15:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:12.146 15:15:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # return 0 00:11:12.146 15:15:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:12.146 15:15:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:12.146 15:15:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.146 15:15:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.146 BaseBdev1_malloc 00:11:12.146 15:15:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.146 15:15:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:12.146 15:15:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.146 15:15:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.146 [2024-11-27 15:15:39.938588] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:12.146 [2024-11-27 15:15:39.938659] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:12.146 [2024-11-27 15:15:39.938695] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:11:12.146 [2024-11-27 15:15:39.938707] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:12.146 [2024-11-27 15:15:39.941118] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:12.146 [2024-11-27 15:15:39.941256] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:12.146 BaseBdev1 00:11:12.146 15:15:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.146 15:15:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:12.146 15:15:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:12.146 15:15:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.146 15:15:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.146 BaseBdev2_malloc 00:11:12.146 15:15:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.146 15:15:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:11:12.146 15:15:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.146 15:15:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.146 [2024-11-27 15:15:39.967491] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:11:12.146 [2024-11-27 15:15:39.967557] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:12.146 [2024-11-27 15:15:39.967579] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:11:12.146 [2024-11-27 15:15:39.967589] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:12.146 [2024-11-27 15:15:39.969748] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:12.146 [2024-11-27 15:15:39.969787] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:12.146 BaseBdev2 00:11:12.146 15:15:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.146 15:15:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:11:12.146 15:15:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.146 15:15:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.146 spare_malloc 00:11:12.146 15:15:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.146 15:15:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:11:12.146 15:15:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.146 15:15:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.146 spare_delay 00:11:12.146 15:15:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.146 15:15:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:12.146 15:15:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.146 15:15:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.146 [2024-11-27 15:15:40.008490] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:12.146 [2024-11-27 15:15:40.008559] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:12.146 [2024-11-27 15:15:40.008584] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:11:12.146 [2024-11-27 15:15:40.008594] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:12.146 [2024-11-27 15:15:40.010944] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:12.146 [2024-11-27 15:15:40.010982] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:12.146 spare 00:11:12.146 15:15:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.146 15:15:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:11:12.147 15:15:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.147 15:15:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.147 [2024-11-27 15:15:40.020489] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:12.147 [2024-11-27 15:15:40.022284] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:12.147 [2024-11-27 15:15:40.022455] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:11:12.147 [2024-11-27 15:15:40.022469] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:11:12.147 [2024-11-27 15:15:40.022709] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:11:12.147 [2024-11-27 15:15:40.022841] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:11:12.147 [2024-11-27 15:15:40.022856] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:11:12.147 [2024-11-27 15:15:40.022990] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:12.147 15:15:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.147 15:15:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:12.147 15:15:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:12.147 15:15:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:12.147 15:15:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:12.147 15:15:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:12.147 15:15:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:12.147 15:15:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:12.147 15:15:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:12.147 15:15:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:12.147 15:15:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:12.147 15:15:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:12.147 15:15:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:12.147 15:15:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.147 15:15:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.147 15:15:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.147 15:15:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:12.147 "name": "raid_bdev1", 00:11:12.147 "uuid": "07f86d65-eb86-4343-ac78-e2ee8641af97", 00:11:12.147 "strip_size_kb": 0, 00:11:12.147 "state": "online", 00:11:12.147 "raid_level": "raid1", 00:11:12.147 "superblock": false, 00:11:12.147 "num_base_bdevs": 2, 00:11:12.147 "num_base_bdevs_discovered": 2, 00:11:12.147 "num_base_bdevs_operational": 2, 00:11:12.147 "base_bdevs_list": [ 00:11:12.147 { 00:11:12.147 "name": "BaseBdev1", 00:11:12.147 "uuid": "1589149e-4951-51db-8bfe-948b35ee322e", 00:11:12.147 "is_configured": true, 00:11:12.147 "data_offset": 0, 00:11:12.147 "data_size": 65536 00:11:12.147 }, 00:11:12.147 { 00:11:12.147 "name": "BaseBdev2", 00:11:12.147 "uuid": "4b5f5d34-1f8f-564d-ad06-7ce8f59151d9", 00:11:12.147 "is_configured": true, 00:11:12.147 "data_offset": 0, 00:11:12.147 "data_size": 65536 00:11:12.147 } 00:11:12.147 ] 00:11:12.147 }' 00:11:12.147 15:15:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:12.147 15:15:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.407 15:15:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:11:12.407 15:15:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:12.407 15:15:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.407 15:15:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.407 [2024-11-27 15:15:40.444105] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:12.407 15:15:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.407 15:15:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:11:12.407 15:15:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:11:12.407 15:15:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:12.407 15:15:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.407 15:15:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.407 15:15:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.407 15:15:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:11:12.407 15:15:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:11:12.407 15:15:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:11:12.407 15:15:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:11:12.407 15:15:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:11:12.407 15:15:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:12.407 15:15:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:11:12.407 15:15:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:12.407 15:15:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:11:12.407 15:15:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:12.407 15:15:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:11:12.407 15:15:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:12.407 15:15:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:12.407 15:15:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:11:12.666 [2024-11-27 15:15:40.719600] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:11:12.666 /dev/nbd0 00:11:12.666 15:15:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:12.927 15:15:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:12.927 15:15:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:11:12.927 15:15:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:11:12.927 15:15:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:12.927 15:15:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:12.927 15:15:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:11:12.927 15:15:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:11:12.927 15:15:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:12.927 15:15:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:12.927 15:15:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:12.927 1+0 records in 00:11:12.927 1+0 records out 00:11:12.927 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000370823 s, 11.0 MB/s 00:11:12.927 15:15:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:12.927 15:15:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:11:12.927 15:15:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:12.927 15:15:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:12.927 15:15:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:11:12.927 15:15:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:12.927 15:15:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:12.927 15:15:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:11:12.927 15:15:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:11:12.927 15:15:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=65536 oflag=direct 00:11:18.212 65536+0 records in 00:11:18.212 65536+0 records out 00:11:18.212 33554432 bytes (34 MB, 32 MiB) copied, 4.92772 s, 6.8 MB/s 00:11:18.212 15:15:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:11:18.212 15:15:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:18.213 15:15:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:11:18.213 15:15:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:18.213 15:15:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:11:18.213 15:15:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:18.213 15:15:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:18.213 [2024-11-27 15:15:45.932599] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:18.213 15:15:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:18.213 15:15:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:18.213 15:15:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:18.213 15:15:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:18.213 15:15:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:18.213 15:15:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:18.213 15:15:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:11:18.213 15:15:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:11:18.213 15:15:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:11:18.213 15:15:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:18.213 15:15:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:18.213 [2024-11-27 15:15:45.968600] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:18.213 15:15:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:18.213 15:15:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:18.213 15:15:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:18.213 15:15:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:18.213 15:15:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:18.213 15:15:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:18.213 15:15:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:18.213 15:15:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:18.213 15:15:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:18.213 15:15:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:18.213 15:15:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:18.213 15:15:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:18.213 15:15:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:18.213 15:15:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:18.213 15:15:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:18.213 15:15:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:18.213 15:15:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:18.213 "name": "raid_bdev1", 00:11:18.213 "uuid": "07f86d65-eb86-4343-ac78-e2ee8641af97", 00:11:18.213 "strip_size_kb": 0, 00:11:18.213 "state": "online", 00:11:18.213 "raid_level": "raid1", 00:11:18.213 "superblock": false, 00:11:18.213 "num_base_bdevs": 2, 00:11:18.213 "num_base_bdevs_discovered": 1, 00:11:18.213 "num_base_bdevs_operational": 1, 00:11:18.213 "base_bdevs_list": [ 00:11:18.213 { 00:11:18.213 "name": null, 00:11:18.213 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:18.213 "is_configured": false, 00:11:18.213 "data_offset": 0, 00:11:18.213 "data_size": 65536 00:11:18.213 }, 00:11:18.213 { 00:11:18.213 "name": "BaseBdev2", 00:11:18.213 "uuid": "4b5f5d34-1f8f-564d-ad06-7ce8f59151d9", 00:11:18.213 "is_configured": true, 00:11:18.213 "data_offset": 0, 00:11:18.213 "data_size": 65536 00:11:18.213 } 00:11:18.213 ] 00:11:18.213 }' 00:11:18.213 15:15:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:18.213 15:15:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:18.472 15:15:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:18.472 15:15:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:18.472 15:15:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:18.472 [2024-11-27 15:15:46.419863] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:18.472 [2024-11-27 15:15:46.424860] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d09a30 00:11:18.472 15:15:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:18.472 15:15:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:11:18.472 [2024-11-27 15:15:46.426790] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:19.410 15:15:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:19.410 15:15:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:19.410 15:15:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:19.410 15:15:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:19.410 15:15:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:19.410 15:15:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:19.410 15:15:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:19.410 15:15:47 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:19.410 15:15:47 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:19.410 15:15:47 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:19.410 15:15:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:19.410 "name": "raid_bdev1", 00:11:19.410 "uuid": "07f86d65-eb86-4343-ac78-e2ee8641af97", 00:11:19.410 "strip_size_kb": 0, 00:11:19.410 "state": "online", 00:11:19.410 "raid_level": "raid1", 00:11:19.410 "superblock": false, 00:11:19.410 "num_base_bdevs": 2, 00:11:19.410 "num_base_bdevs_discovered": 2, 00:11:19.410 "num_base_bdevs_operational": 2, 00:11:19.410 "process": { 00:11:19.410 "type": "rebuild", 00:11:19.410 "target": "spare", 00:11:19.410 "progress": { 00:11:19.410 "blocks": 20480, 00:11:19.410 "percent": 31 00:11:19.410 } 00:11:19.410 }, 00:11:19.410 "base_bdevs_list": [ 00:11:19.410 { 00:11:19.410 "name": "spare", 00:11:19.410 "uuid": "dc2847f2-6699-5e84-b5eb-a14d792a3e81", 00:11:19.410 "is_configured": true, 00:11:19.410 "data_offset": 0, 00:11:19.410 "data_size": 65536 00:11:19.410 }, 00:11:19.410 { 00:11:19.410 "name": "BaseBdev2", 00:11:19.410 "uuid": "4b5f5d34-1f8f-564d-ad06-7ce8f59151d9", 00:11:19.410 "is_configured": true, 00:11:19.410 "data_offset": 0, 00:11:19.410 "data_size": 65536 00:11:19.410 } 00:11:19.410 ] 00:11:19.410 }' 00:11:19.410 15:15:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:19.410 15:15:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:19.410 15:15:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:19.669 15:15:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:19.669 15:15:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:19.669 15:15:47 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:19.669 15:15:47 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:19.669 [2024-11-27 15:15:47.567708] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:19.669 [2024-11-27 15:15:47.632122] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:19.669 [2024-11-27 15:15:47.632198] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:19.669 [2024-11-27 15:15:47.632219] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:19.669 [2024-11-27 15:15:47.632227] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:19.669 15:15:47 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:19.669 15:15:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:19.669 15:15:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:19.669 15:15:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:19.669 15:15:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:19.669 15:15:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:19.669 15:15:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:19.669 15:15:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:19.669 15:15:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:19.669 15:15:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:19.669 15:15:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:19.669 15:15:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:19.669 15:15:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:19.669 15:15:47 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:19.669 15:15:47 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:19.669 15:15:47 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:19.669 15:15:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:19.669 "name": "raid_bdev1", 00:11:19.669 "uuid": "07f86d65-eb86-4343-ac78-e2ee8641af97", 00:11:19.669 "strip_size_kb": 0, 00:11:19.669 "state": "online", 00:11:19.669 "raid_level": "raid1", 00:11:19.669 "superblock": false, 00:11:19.669 "num_base_bdevs": 2, 00:11:19.669 "num_base_bdevs_discovered": 1, 00:11:19.669 "num_base_bdevs_operational": 1, 00:11:19.669 "base_bdevs_list": [ 00:11:19.669 { 00:11:19.669 "name": null, 00:11:19.669 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:19.669 "is_configured": false, 00:11:19.669 "data_offset": 0, 00:11:19.669 "data_size": 65536 00:11:19.669 }, 00:11:19.669 { 00:11:19.669 "name": "BaseBdev2", 00:11:19.669 "uuid": "4b5f5d34-1f8f-564d-ad06-7ce8f59151d9", 00:11:19.669 "is_configured": true, 00:11:19.669 "data_offset": 0, 00:11:19.669 "data_size": 65536 00:11:19.669 } 00:11:19.669 ] 00:11:19.669 }' 00:11:19.669 15:15:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:19.669 15:15:47 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.236 15:15:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:20.236 15:15:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:20.236 15:15:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:20.236 15:15:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:20.236 15:15:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:20.236 15:15:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:20.236 15:15:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:20.236 15:15:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:20.236 15:15:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.236 15:15:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.236 15:15:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:20.236 "name": "raid_bdev1", 00:11:20.236 "uuid": "07f86d65-eb86-4343-ac78-e2ee8641af97", 00:11:20.236 "strip_size_kb": 0, 00:11:20.236 "state": "online", 00:11:20.236 "raid_level": "raid1", 00:11:20.236 "superblock": false, 00:11:20.236 "num_base_bdevs": 2, 00:11:20.236 "num_base_bdevs_discovered": 1, 00:11:20.236 "num_base_bdevs_operational": 1, 00:11:20.236 "base_bdevs_list": [ 00:11:20.236 { 00:11:20.236 "name": null, 00:11:20.236 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:20.236 "is_configured": false, 00:11:20.236 "data_offset": 0, 00:11:20.236 "data_size": 65536 00:11:20.236 }, 00:11:20.236 { 00:11:20.236 "name": "BaseBdev2", 00:11:20.236 "uuid": "4b5f5d34-1f8f-564d-ad06-7ce8f59151d9", 00:11:20.236 "is_configured": true, 00:11:20.236 "data_offset": 0, 00:11:20.236 "data_size": 65536 00:11:20.236 } 00:11:20.236 ] 00:11:20.236 }' 00:11:20.236 15:15:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:20.236 15:15:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:20.236 15:15:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:20.236 15:15:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:20.236 15:15:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:20.236 15:15:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:20.236 15:15:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.236 [2024-11-27 15:15:48.256313] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:20.236 [2024-11-27 15:15:48.261395] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d09b00 00:11:20.236 15:15:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.236 15:15:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:11:20.236 [2024-11-27 15:15:48.263296] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:21.175 15:15:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:21.175 15:15:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:21.175 15:15:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:21.175 15:15:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:21.175 15:15:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:21.175 15:15:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:21.175 15:15:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:21.175 15:15:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.175 15:15:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.435 15:15:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.435 15:15:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:21.435 "name": "raid_bdev1", 00:11:21.435 "uuid": "07f86d65-eb86-4343-ac78-e2ee8641af97", 00:11:21.435 "strip_size_kb": 0, 00:11:21.435 "state": "online", 00:11:21.435 "raid_level": "raid1", 00:11:21.435 "superblock": false, 00:11:21.435 "num_base_bdevs": 2, 00:11:21.435 "num_base_bdevs_discovered": 2, 00:11:21.435 "num_base_bdevs_operational": 2, 00:11:21.435 "process": { 00:11:21.435 "type": "rebuild", 00:11:21.435 "target": "spare", 00:11:21.435 "progress": { 00:11:21.435 "blocks": 20480, 00:11:21.435 "percent": 31 00:11:21.435 } 00:11:21.435 }, 00:11:21.435 "base_bdevs_list": [ 00:11:21.435 { 00:11:21.435 "name": "spare", 00:11:21.435 "uuid": "dc2847f2-6699-5e84-b5eb-a14d792a3e81", 00:11:21.435 "is_configured": true, 00:11:21.435 "data_offset": 0, 00:11:21.435 "data_size": 65536 00:11:21.435 }, 00:11:21.435 { 00:11:21.435 "name": "BaseBdev2", 00:11:21.435 "uuid": "4b5f5d34-1f8f-564d-ad06-7ce8f59151d9", 00:11:21.435 "is_configured": true, 00:11:21.435 "data_offset": 0, 00:11:21.435 "data_size": 65536 00:11:21.435 } 00:11:21.435 ] 00:11:21.435 }' 00:11:21.435 15:15:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:21.435 15:15:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:21.435 15:15:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:21.435 15:15:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:21.435 15:15:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:11:21.435 15:15:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:11:21.435 15:15:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:11:21.435 15:15:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:11:21.435 15:15:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=290 00:11:21.435 15:15:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:21.435 15:15:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:21.435 15:15:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:21.435 15:15:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:21.435 15:15:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:21.435 15:15:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:21.435 15:15:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:21.435 15:15:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:21.435 15:15:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.435 15:15:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.435 15:15:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.435 15:15:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:21.435 "name": "raid_bdev1", 00:11:21.435 "uuid": "07f86d65-eb86-4343-ac78-e2ee8641af97", 00:11:21.435 "strip_size_kb": 0, 00:11:21.435 "state": "online", 00:11:21.435 "raid_level": "raid1", 00:11:21.435 "superblock": false, 00:11:21.435 "num_base_bdevs": 2, 00:11:21.435 "num_base_bdevs_discovered": 2, 00:11:21.435 "num_base_bdevs_operational": 2, 00:11:21.435 "process": { 00:11:21.435 "type": "rebuild", 00:11:21.435 "target": "spare", 00:11:21.435 "progress": { 00:11:21.435 "blocks": 22528, 00:11:21.435 "percent": 34 00:11:21.435 } 00:11:21.435 }, 00:11:21.435 "base_bdevs_list": [ 00:11:21.435 { 00:11:21.435 "name": "spare", 00:11:21.435 "uuid": "dc2847f2-6699-5e84-b5eb-a14d792a3e81", 00:11:21.435 "is_configured": true, 00:11:21.435 "data_offset": 0, 00:11:21.435 "data_size": 65536 00:11:21.435 }, 00:11:21.435 { 00:11:21.435 "name": "BaseBdev2", 00:11:21.435 "uuid": "4b5f5d34-1f8f-564d-ad06-7ce8f59151d9", 00:11:21.435 "is_configured": true, 00:11:21.435 "data_offset": 0, 00:11:21.435 "data_size": 65536 00:11:21.435 } 00:11:21.435 ] 00:11:21.435 }' 00:11:21.435 15:15:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:21.435 15:15:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:21.435 15:15:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:21.695 15:15:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:21.695 15:15:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:22.635 15:15:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:22.635 15:15:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:22.635 15:15:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:22.635 15:15:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:22.635 15:15:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:22.635 15:15:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:22.635 15:15:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:22.635 15:15:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:22.635 15:15:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:22.635 15:15:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.635 15:15:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:22.635 15:15:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:22.635 "name": "raid_bdev1", 00:11:22.635 "uuid": "07f86d65-eb86-4343-ac78-e2ee8641af97", 00:11:22.635 "strip_size_kb": 0, 00:11:22.635 "state": "online", 00:11:22.635 "raid_level": "raid1", 00:11:22.635 "superblock": false, 00:11:22.635 "num_base_bdevs": 2, 00:11:22.635 "num_base_bdevs_discovered": 2, 00:11:22.635 "num_base_bdevs_operational": 2, 00:11:22.635 "process": { 00:11:22.635 "type": "rebuild", 00:11:22.635 "target": "spare", 00:11:22.635 "progress": { 00:11:22.635 "blocks": 47104, 00:11:22.635 "percent": 71 00:11:22.635 } 00:11:22.635 }, 00:11:22.635 "base_bdevs_list": [ 00:11:22.635 { 00:11:22.635 "name": "spare", 00:11:22.635 "uuid": "dc2847f2-6699-5e84-b5eb-a14d792a3e81", 00:11:22.635 "is_configured": true, 00:11:22.635 "data_offset": 0, 00:11:22.635 "data_size": 65536 00:11:22.635 }, 00:11:22.635 { 00:11:22.635 "name": "BaseBdev2", 00:11:22.635 "uuid": "4b5f5d34-1f8f-564d-ad06-7ce8f59151d9", 00:11:22.635 "is_configured": true, 00:11:22.635 "data_offset": 0, 00:11:22.635 "data_size": 65536 00:11:22.635 } 00:11:22.635 ] 00:11:22.635 }' 00:11:22.635 15:15:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:22.635 15:15:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:22.635 15:15:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:22.635 15:15:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:22.635 15:15:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:23.574 [2024-11-27 15:15:51.476341] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:11:23.574 [2024-11-27 15:15:51.476533] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:11:23.574 [2024-11-27 15:15:51.476583] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:23.833 15:15:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:23.833 15:15:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:23.834 15:15:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:23.834 15:15:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:23.834 15:15:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:23.834 15:15:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:23.834 15:15:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:23.834 15:15:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:23.834 15:15:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.834 15:15:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.834 15:15:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.834 15:15:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:23.834 "name": "raid_bdev1", 00:11:23.834 "uuid": "07f86d65-eb86-4343-ac78-e2ee8641af97", 00:11:23.834 "strip_size_kb": 0, 00:11:23.834 "state": "online", 00:11:23.834 "raid_level": "raid1", 00:11:23.834 "superblock": false, 00:11:23.834 "num_base_bdevs": 2, 00:11:23.834 "num_base_bdevs_discovered": 2, 00:11:23.834 "num_base_bdevs_operational": 2, 00:11:23.834 "base_bdevs_list": [ 00:11:23.834 { 00:11:23.834 "name": "spare", 00:11:23.834 "uuid": "dc2847f2-6699-5e84-b5eb-a14d792a3e81", 00:11:23.834 "is_configured": true, 00:11:23.834 "data_offset": 0, 00:11:23.834 "data_size": 65536 00:11:23.834 }, 00:11:23.834 { 00:11:23.834 "name": "BaseBdev2", 00:11:23.834 "uuid": "4b5f5d34-1f8f-564d-ad06-7ce8f59151d9", 00:11:23.834 "is_configured": true, 00:11:23.834 "data_offset": 0, 00:11:23.834 "data_size": 65536 00:11:23.834 } 00:11:23.834 ] 00:11:23.834 }' 00:11:23.834 15:15:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:23.834 15:15:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:11:23.834 15:15:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:23.834 15:15:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:11:23.834 15:15:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:11:23.834 15:15:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:23.834 15:15:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:23.834 15:15:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:23.834 15:15:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:23.834 15:15:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:23.834 15:15:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:23.834 15:15:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:23.834 15:15:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.834 15:15:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.834 15:15:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.834 15:15:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:23.834 "name": "raid_bdev1", 00:11:23.834 "uuid": "07f86d65-eb86-4343-ac78-e2ee8641af97", 00:11:23.834 "strip_size_kb": 0, 00:11:23.834 "state": "online", 00:11:23.834 "raid_level": "raid1", 00:11:23.834 "superblock": false, 00:11:23.834 "num_base_bdevs": 2, 00:11:23.834 "num_base_bdevs_discovered": 2, 00:11:23.834 "num_base_bdevs_operational": 2, 00:11:23.834 "base_bdevs_list": [ 00:11:23.834 { 00:11:23.834 "name": "spare", 00:11:23.834 "uuid": "dc2847f2-6699-5e84-b5eb-a14d792a3e81", 00:11:23.834 "is_configured": true, 00:11:23.834 "data_offset": 0, 00:11:23.834 "data_size": 65536 00:11:23.834 }, 00:11:23.834 { 00:11:23.834 "name": "BaseBdev2", 00:11:23.834 "uuid": "4b5f5d34-1f8f-564d-ad06-7ce8f59151d9", 00:11:23.834 "is_configured": true, 00:11:23.834 "data_offset": 0, 00:11:23.834 "data_size": 65536 00:11:23.834 } 00:11:23.834 ] 00:11:23.834 }' 00:11:23.834 15:15:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:24.095 15:15:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:24.095 15:15:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:24.095 15:15:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:24.095 15:15:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:24.095 15:15:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:24.095 15:15:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:24.095 15:15:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:24.095 15:15:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:24.095 15:15:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:24.095 15:15:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:24.095 15:15:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:24.095 15:15:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:24.095 15:15:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:24.095 15:15:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:24.095 15:15:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:24.095 15:15:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.095 15:15:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.095 15:15:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.095 15:15:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:24.095 "name": "raid_bdev1", 00:11:24.095 "uuid": "07f86d65-eb86-4343-ac78-e2ee8641af97", 00:11:24.095 "strip_size_kb": 0, 00:11:24.095 "state": "online", 00:11:24.095 "raid_level": "raid1", 00:11:24.095 "superblock": false, 00:11:24.095 "num_base_bdevs": 2, 00:11:24.095 "num_base_bdevs_discovered": 2, 00:11:24.095 "num_base_bdevs_operational": 2, 00:11:24.095 "base_bdevs_list": [ 00:11:24.095 { 00:11:24.095 "name": "spare", 00:11:24.095 "uuid": "dc2847f2-6699-5e84-b5eb-a14d792a3e81", 00:11:24.095 "is_configured": true, 00:11:24.095 "data_offset": 0, 00:11:24.095 "data_size": 65536 00:11:24.095 }, 00:11:24.095 { 00:11:24.095 "name": "BaseBdev2", 00:11:24.095 "uuid": "4b5f5d34-1f8f-564d-ad06-7ce8f59151d9", 00:11:24.095 "is_configured": true, 00:11:24.095 "data_offset": 0, 00:11:24.095 "data_size": 65536 00:11:24.095 } 00:11:24.095 ] 00:11:24.095 }' 00:11:24.095 15:15:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:24.095 15:15:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.673 15:15:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:24.673 15:15:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.673 15:15:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.673 [2024-11-27 15:15:52.483797] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:24.673 [2024-11-27 15:15:52.483890] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:24.673 [2024-11-27 15:15:52.483994] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:24.673 [2024-11-27 15:15:52.484063] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:24.673 [2024-11-27 15:15:52.484083] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:11:24.673 15:15:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.673 15:15:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:11:24.673 15:15:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:24.673 15:15:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.673 15:15:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.673 15:15:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.673 15:15:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:11:24.673 15:15:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:11:24.674 15:15:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:11:24.674 15:15:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:11:24.674 15:15:52 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:24.674 15:15:52 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:11:24.674 15:15:52 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:24.674 15:15:52 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:11:24.674 15:15:52 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:24.674 15:15:52 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:11:24.674 15:15:52 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:24.674 15:15:52 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:24.674 15:15:52 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:11:24.674 /dev/nbd0 00:11:24.674 15:15:52 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:24.674 15:15:52 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:24.674 15:15:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:11:24.674 15:15:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:11:24.674 15:15:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:24.674 15:15:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:24.674 15:15:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:11:24.674 15:15:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:11:24.674 15:15:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:24.674 15:15:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:24.674 15:15:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:24.674 1+0 records in 00:11:24.674 1+0 records out 00:11:24.674 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000448322 s, 9.1 MB/s 00:11:24.674 15:15:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:24.934 15:15:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:11:24.934 15:15:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:24.934 15:15:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:24.934 15:15:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:11:24.934 15:15:52 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:24.934 15:15:52 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:24.934 15:15:52 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:11:24.934 /dev/nbd1 00:11:24.934 15:15:53 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:11:24.934 15:15:53 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:11:24.934 15:15:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:11:24.934 15:15:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:11:24.934 15:15:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:24.934 15:15:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:24.934 15:15:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:11:24.934 15:15:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:11:24.934 15:15:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:24.934 15:15:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:24.934 15:15:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:24.934 1+0 records in 00:11:24.934 1+0 records out 00:11:24.934 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000246001 s, 16.7 MB/s 00:11:24.934 15:15:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:24.934 15:15:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:11:24.934 15:15:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:24.934 15:15:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:24.934 15:15:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:11:24.934 15:15:53 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:24.934 15:15:53 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:24.934 15:15:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:11:25.192 15:15:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:11:25.192 15:15:53 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:25.192 15:15:53 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:11:25.192 15:15:53 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:25.192 15:15:53 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:11:25.192 15:15:53 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:25.192 15:15:53 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:25.450 15:15:53 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:25.450 15:15:53 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:25.450 15:15:53 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:25.450 15:15:53 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:25.450 15:15:53 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:25.450 15:15:53 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:25.450 15:15:53 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:11:25.450 15:15:53 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:11:25.450 15:15:53 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:25.450 15:15:53 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:11:25.450 15:15:53 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:11:25.450 15:15:53 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:11:25.450 15:15:53 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:11:25.450 15:15:53 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:25.450 15:15:53 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:25.450 15:15:53 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:11:25.450 15:15:53 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:11:25.450 15:15:53 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:11:25.450 15:15:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:11:25.450 15:15:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 86158 00:11:25.450 15:15:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@954 -- # '[' -z 86158 ']' 00:11:25.450 15:15:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@958 -- # kill -0 86158 00:11:25.450 15:15:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@959 -- # uname 00:11:25.709 15:15:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:25.709 15:15:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 86158 00:11:25.709 killing process with pid 86158 00:11:25.709 Received shutdown signal, test time was about 60.000000 seconds 00:11:25.709 00:11:25.709 Latency(us) 00:11:25.709 [2024-11-27T15:15:53.816Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:25.709 [2024-11-27T15:15:53.816Z] =================================================================================================================== 00:11:25.709 [2024-11-27T15:15:53.816Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:11:25.709 15:15:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:25.709 15:15:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:25.709 15:15:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 86158' 00:11:25.709 15:15:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@973 -- # kill 86158 00:11:25.709 [2024-11-27 15:15:53.591425] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:25.709 15:15:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@978 -- # wait 86158 00:11:25.709 [2024-11-27 15:15:53.622551] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:11:25.969 00:11:25.969 real 0m14.900s 00:11:25.969 user 0m16.647s 00:11:25.969 sys 0m3.369s 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.969 ************************************ 00:11:25.969 END TEST raid_rebuild_test 00:11:25.969 ************************************ 00:11:25.969 15:15:53 bdev_raid -- bdev/bdev_raid.sh@979 -- # run_test raid_rebuild_test_sb raid_rebuild_test raid1 2 true false true 00:11:25.969 15:15:53 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:11:25.969 15:15:53 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:25.969 15:15:53 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:25.969 ************************************ 00:11:25.969 START TEST raid_rebuild_test_sb 00:11:25.969 ************************************ 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true false true 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=86572 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 86572 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@835 -- # '[' -z 86572 ']' 00:11:25.969 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:25.969 15:15:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:25.969 [2024-11-27 15:15:54.003119] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:11:25.969 I/O size of 3145728 is greater than zero copy threshold (65536). 00:11:25.969 Zero copy mechanism will not be used. 00:11:25.969 [2024-11-27 15:15:54.003325] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid86572 ] 00:11:26.228 [2024-11-27 15:15:54.171958] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:26.228 [2024-11-27 15:15:54.198520] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:26.228 [2024-11-27 15:15:54.242069] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:26.228 [2024-11-27 15:15:54.242108] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:26.796 15:15:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:26.796 15:15:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # return 0 00:11:26.796 15:15:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:26.796 15:15:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:26.796 15:15:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:26.796 15:15:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:26.796 BaseBdev1_malloc 00:11:26.796 15:15:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:26.796 15:15:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:26.796 15:15:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:26.796 15:15:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:26.796 [2024-11-27 15:15:54.882524] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:26.797 [2024-11-27 15:15:54.882584] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:26.797 [2024-11-27 15:15:54.882611] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:11:26.797 [2024-11-27 15:15:54.882637] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:26.797 [2024-11-27 15:15:54.884809] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:26.797 [2024-11-27 15:15:54.884847] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:26.797 BaseBdev1 00:11:26.797 15:15:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:26.797 15:15:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:26.797 15:15:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:26.797 15:15:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:26.797 15:15:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:27.056 BaseBdev2_malloc 00:11:27.056 15:15:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:27.056 15:15:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:11:27.056 15:15:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:27.056 15:15:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:27.056 [2024-11-27 15:15:54.911735] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:11:27.056 [2024-11-27 15:15:54.911794] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:27.056 [2024-11-27 15:15:54.911819] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:11:27.056 [2024-11-27 15:15:54.911829] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:27.056 [2024-11-27 15:15:54.914029] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:27.056 [2024-11-27 15:15:54.914110] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:27.056 BaseBdev2 00:11:27.056 15:15:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:27.056 15:15:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:11:27.056 15:15:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:27.056 15:15:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:27.056 spare_malloc 00:11:27.056 15:15:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:27.056 15:15:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:11:27.056 15:15:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:27.056 15:15:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:27.056 spare_delay 00:11:27.056 15:15:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:27.056 15:15:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:27.056 15:15:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:27.056 15:15:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:27.056 [2024-11-27 15:15:54.952990] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:27.056 [2024-11-27 15:15:54.953102] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:27.056 [2024-11-27 15:15:54.953129] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:11:27.056 [2024-11-27 15:15:54.953139] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:27.056 [2024-11-27 15:15:54.955269] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:27.056 [2024-11-27 15:15:54.955308] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:27.056 spare 00:11:27.056 15:15:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:27.056 15:15:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:11:27.056 15:15:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:27.056 15:15:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:27.056 [2024-11-27 15:15:54.965016] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:27.057 [2024-11-27 15:15:54.966816] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:27.057 [2024-11-27 15:15:54.967011] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:11:27.057 [2024-11-27 15:15:54.967025] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:27.057 [2024-11-27 15:15:54.967318] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:11:27.057 [2024-11-27 15:15:54.967491] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:11:27.057 [2024-11-27 15:15:54.967505] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:11:27.057 [2024-11-27 15:15:54.967642] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:27.057 15:15:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:27.057 15:15:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:27.057 15:15:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:27.057 15:15:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:27.057 15:15:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:27.057 15:15:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:27.057 15:15:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:27.057 15:15:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:27.057 15:15:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:27.057 15:15:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:27.057 15:15:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:27.057 15:15:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:27.057 15:15:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:27.057 15:15:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:27.057 15:15:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:27.057 15:15:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:27.057 15:15:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:27.057 "name": "raid_bdev1", 00:11:27.057 "uuid": "d9fe91ae-3d72-44a3-83a7-3a94d10441b0", 00:11:27.057 "strip_size_kb": 0, 00:11:27.057 "state": "online", 00:11:27.057 "raid_level": "raid1", 00:11:27.057 "superblock": true, 00:11:27.057 "num_base_bdevs": 2, 00:11:27.057 "num_base_bdevs_discovered": 2, 00:11:27.057 "num_base_bdevs_operational": 2, 00:11:27.057 "base_bdevs_list": [ 00:11:27.057 { 00:11:27.057 "name": "BaseBdev1", 00:11:27.057 "uuid": "d43b0a36-7a47-5baa-b6c0-cd9846f05f86", 00:11:27.057 "is_configured": true, 00:11:27.057 "data_offset": 2048, 00:11:27.057 "data_size": 63488 00:11:27.057 }, 00:11:27.057 { 00:11:27.057 "name": "BaseBdev2", 00:11:27.057 "uuid": "284114bb-d82a-54f3-a4eb-496e1224678c", 00:11:27.057 "is_configured": true, 00:11:27.057 "data_offset": 2048, 00:11:27.057 "data_size": 63488 00:11:27.057 } 00:11:27.057 ] 00:11:27.057 }' 00:11:27.057 15:15:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:27.057 15:15:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:27.316 15:15:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:27.316 15:15:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:11:27.316 15:15:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:27.316 15:15:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:27.316 [2024-11-27 15:15:55.420540] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:27.576 15:15:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:27.576 15:15:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:11:27.576 15:15:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:11:27.576 15:15:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:27.576 15:15:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:27.576 15:15:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:27.576 15:15:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:27.576 15:15:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:11:27.576 15:15:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:11:27.576 15:15:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:11:27.576 15:15:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:11:27.576 15:15:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:11:27.576 15:15:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:27.576 15:15:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:11:27.576 15:15:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:27.576 15:15:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:11:27.576 15:15:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:27.576 15:15:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:11:27.576 15:15:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:27.576 15:15:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:27.576 15:15:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:11:27.836 [2024-11-27 15:15:55.695780] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:11:27.836 /dev/nbd0 00:11:27.836 15:15:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:27.836 15:15:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:27.836 15:15:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:11:27.836 15:15:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:11:27.836 15:15:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:27.836 15:15:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:27.836 15:15:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:11:27.836 15:15:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:11:27.836 15:15:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:27.836 15:15:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:27.836 15:15:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:27.836 1+0 records in 00:11:27.836 1+0 records out 00:11:27.836 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00037898 s, 10.8 MB/s 00:11:27.836 15:15:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:27.836 15:15:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:11:27.836 15:15:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:27.836 15:15:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:27.836 15:15:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:11:27.836 15:15:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:27.836 15:15:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:27.836 15:15:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:11:27.836 15:15:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:11:27.836 15:15:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=63488 oflag=direct 00:11:33.112 63488+0 records in 00:11:33.112 63488+0 records out 00:11:33.112 32505856 bytes (33 MB, 31 MiB) copied, 4.43314 s, 7.3 MB/s 00:11:33.112 15:16:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:11:33.112 15:16:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:33.112 15:16:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:11:33.112 15:16:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:33.112 15:16:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:11:33.112 15:16:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:33.112 15:16:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:33.112 [2024-11-27 15:16:00.435106] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:33.112 15:16:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:33.112 15:16:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:33.112 15:16:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:33.112 15:16:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:33.112 15:16:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:33.112 15:16:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:33.112 15:16:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:11:33.112 15:16:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:11:33.112 15:16:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:11:33.112 15:16:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.112 15:16:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:33.112 [2024-11-27 15:16:00.471154] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:33.112 15:16:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.112 15:16:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:33.113 15:16:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:33.113 15:16:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:33.113 15:16:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:33.113 15:16:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:33.113 15:16:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:33.113 15:16:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:33.113 15:16:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:33.113 15:16:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:33.113 15:16:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:33.113 15:16:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:33.113 15:16:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.113 15:16:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:33.113 15:16:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:33.113 15:16:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.113 15:16:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:33.113 "name": "raid_bdev1", 00:11:33.113 "uuid": "d9fe91ae-3d72-44a3-83a7-3a94d10441b0", 00:11:33.113 "strip_size_kb": 0, 00:11:33.113 "state": "online", 00:11:33.113 "raid_level": "raid1", 00:11:33.113 "superblock": true, 00:11:33.113 "num_base_bdevs": 2, 00:11:33.113 "num_base_bdevs_discovered": 1, 00:11:33.113 "num_base_bdevs_operational": 1, 00:11:33.113 "base_bdevs_list": [ 00:11:33.113 { 00:11:33.113 "name": null, 00:11:33.113 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:33.113 "is_configured": false, 00:11:33.113 "data_offset": 0, 00:11:33.113 "data_size": 63488 00:11:33.113 }, 00:11:33.113 { 00:11:33.113 "name": "BaseBdev2", 00:11:33.113 "uuid": "284114bb-d82a-54f3-a4eb-496e1224678c", 00:11:33.113 "is_configured": true, 00:11:33.113 "data_offset": 2048, 00:11:33.113 "data_size": 63488 00:11:33.113 } 00:11:33.113 ] 00:11:33.113 }' 00:11:33.113 15:16:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:33.113 15:16:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:33.113 15:16:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:33.113 15:16:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.113 15:16:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:33.113 [2024-11-27 15:16:00.950408] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:33.113 [2024-11-27 15:16:00.959562] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000ca31c0 00:11:33.113 15:16:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.113 15:16:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:11:33.113 [2024-11-27 15:16:00.961957] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:34.051 15:16:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:34.051 15:16:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:34.051 15:16:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:34.051 15:16:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:34.051 15:16:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:34.051 15:16:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:34.051 15:16:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:34.051 15:16:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.051 15:16:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:34.051 15:16:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.051 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:34.051 "name": "raid_bdev1", 00:11:34.051 "uuid": "d9fe91ae-3d72-44a3-83a7-3a94d10441b0", 00:11:34.051 "strip_size_kb": 0, 00:11:34.051 "state": "online", 00:11:34.051 "raid_level": "raid1", 00:11:34.051 "superblock": true, 00:11:34.051 "num_base_bdevs": 2, 00:11:34.051 "num_base_bdevs_discovered": 2, 00:11:34.051 "num_base_bdevs_operational": 2, 00:11:34.051 "process": { 00:11:34.051 "type": "rebuild", 00:11:34.051 "target": "spare", 00:11:34.051 "progress": { 00:11:34.051 "blocks": 20480, 00:11:34.051 "percent": 32 00:11:34.051 } 00:11:34.051 }, 00:11:34.051 "base_bdevs_list": [ 00:11:34.051 { 00:11:34.051 "name": "spare", 00:11:34.051 "uuid": "01b0005d-400d-53f3-9990-eb6685bc4bd6", 00:11:34.051 "is_configured": true, 00:11:34.051 "data_offset": 2048, 00:11:34.051 "data_size": 63488 00:11:34.051 }, 00:11:34.051 { 00:11:34.051 "name": "BaseBdev2", 00:11:34.051 "uuid": "284114bb-d82a-54f3-a4eb-496e1224678c", 00:11:34.051 "is_configured": true, 00:11:34.051 "data_offset": 2048, 00:11:34.051 "data_size": 63488 00:11:34.051 } 00:11:34.051 ] 00:11:34.051 }' 00:11:34.051 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:34.051 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:34.051 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:34.051 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:34.051 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:34.051 15:16:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.051 15:16:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:34.051 [2024-11-27 15:16:02.121994] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:34.311 [2024-11-27 15:16:02.171711] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:34.311 [2024-11-27 15:16:02.171789] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:34.311 [2024-11-27 15:16:02.171814] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:34.311 [2024-11-27 15:16:02.171825] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:34.311 15:16:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.311 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:34.311 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:34.311 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:34.311 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:34.311 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:34.311 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:34.311 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:34.311 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:34.311 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:34.311 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:34.311 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:34.311 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:34.311 15:16:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.311 15:16:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:34.311 15:16:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.311 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:34.311 "name": "raid_bdev1", 00:11:34.311 "uuid": "d9fe91ae-3d72-44a3-83a7-3a94d10441b0", 00:11:34.311 "strip_size_kb": 0, 00:11:34.311 "state": "online", 00:11:34.311 "raid_level": "raid1", 00:11:34.311 "superblock": true, 00:11:34.311 "num_base_bdevs": 2, 00:11:34.311 "num_base_bdevs_discovered": 1, 00:11:34.311 "num_base_bdevs_operational": 1, 00:11:34.311 "base_bdevs_list": [ 00:11:34.311 { 00:11:34.311 "name": null, 00:11:34.311 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:34.311 "is_configured": false, 00:11:34.311 "data_offset": 0, 00:11:34.311 "data_size": 63488 00:11:34.311 }, 00:11:34.311 { 00:11:34.311 "name": "BaseBdev2", 00:11:34.311 "uuid": "284114bb-d82a-54f3-a4eb-496e1224678c", 00:11:34.311 "is_configured": true, 00:11:34.311 "data_offset": 2048, 00:11:34.311 "data_size": 63488 00:11:34.311 } 00:11:34.311 ] 00:11:34.311 }' 00:11:34.311 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:34.311 15:16:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:34.571 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:34.571 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:34.571 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:34.571 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:34.571 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:34.571 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:34.571 15:16:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.571 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:34.571 15:16:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:34.571 15:16:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.571 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:34.571 "name": "raid_bdev1", 00:11:34.571 "uuid": "d9fe91ae-3d72-44a3-83a7-3a94d10441b0", 00:11:34.571 "strip_size_kb": 0, 00:11:34.571 "state": "online", 00:11:34.571 "raid_level": "raid1", 00:11:34.571 "superblock": true, 00:11:34.571 "num_base_bdevs": 2, 00:11:34.571 "num_base_bdevs_discovered": 1, 00:11:34.571 "num_base_bdevs_operational": 1, 00:11:34.571 "base_bdevs_list": [ 00:11:34.571 { 00:11:34.571 "name": null, 00:11:34.571 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:34.571 "is_configured": false, 00:11:34.571 "data_offset": 0, 00:11:34.571 "data_size": 63488 00:11:34.571 }, 00:11:34.571 { 00:11:34.571 "name": "BaseBdev2", 00:11:34.571 "uuid": "284114bb-d82a-54f3-a4eb-496e1224678c", 00:11:34.571 "is_configured": true, 00:11:34.571 "data_offset": 2048, 00:11:34.571 "data_size": 63488 00:11:34.571 } 00:11:34.571 ] 00:11:34.571 }' 00:11:34.571 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:34.831 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:34.831 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:34.831 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:34.831 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:34.831 15:16:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.831 15:16:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:34.831 [2024-11-27 15:16:02.739850] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:34.831 [2024-11-27 15:16:02.749192] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000ca3290 00:11:34.831 15:16:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.831 15:16:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:11:34.831 [2024-11-27 15:16:02.751665] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:35.769 15:16:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:35.769 15:16:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:35.769 15:16:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:35.769 15:16:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:35.769 15:16:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:35.769 15:16:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:35.769 15:16:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:35.769 15:16:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:35.769 15:16:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:35.769 15:16:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:35.769 15:16:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:35.769 "name": "raid_bdev1", 00:11:35.769 "uuid": "d9fe91ae-3d72-44a3-83a7-3a94d10441b0", 00:11:35.769 "strip_size_kb": 0, 00:11:35.769 "state": "online", 00:11:35.769 "raid_level": "raid1", 00:11:35.769 "superblock": true, 00:11:35.769 "num_base_bdevs": 2, 00:11:35.769 "num_base_bdevs_discovered": 2, 00:11:35.769 "num_base_bdevs_operational": 2, 00:11:35.769 "process": { 00:11:35.769 "type": "rebuild", 00:11:35.769 "target": "spare", 00:11:35.769 "progress": { 00:11:35.769 "blocks": 20480, 00:11:35.769 "percent": 32 00:11:35.769 } 00:11:35.769 }, 00:11:35.769 "base_bdevs_list": [ 00:11:35.769 { 00:11:35.769 "name": "spare", 00:11:35.769 "uuid": "01b0005d-400d-53f3-9990-eb6685bc4bd6", 00:11:35.769 "is_configured": true, 00:11:35.769 "data_offset": 2048, 00:11:35.770 "data_size": 63488 00:11:35.770 }, 00:11:35.770 { 00:11:35.770 "name": "BaseBdev2", 00:11:35.770 "uuid": "284114bb-d82a-54f3-a4eb-496e1224678c", 00:11:35.770 "is_configured": true, 00:11:35.770 "data_offset": 2048, 00:11:35.770 "data_size": 63488 00:11:35.770 } 00:11:35.770 ] 00:11:35.770 }' 00:11:35.770 15:16:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:35.770 15:16:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:35.770 15:16:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:36.030 15:16:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:36.030 15:16:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:11:36.030 15:16:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:11:36.030 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:11:36.030 15:16:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:11:36.030 15:16:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:11:36.030 15:16:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:11:36.030 15:16:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=304 00:11:36.030 15:16:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:36.030 15:16:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:36.030 15:16:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:36.030 15:16:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:36.030 15:16:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:36.030 15:16:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:36.030 15:16:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:36.030 15:16:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:36.030 15:16:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:36.030 15:16:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:36.030 15:16:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:36.030 15:16:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:36.030 "name": "raid_bdev1", 00:11:36.030 "uuid": "d9fe91ae-3d72-44a3-83a7-3a94d10441b0", 00:11:36.030 "strip_size_kb": 0, 00:11:36.030 "state": "online", 00:11:36.030 "raid_level": "raid1", 00:11:36.030 "superblock": true, 00:11:36.030 "num_base_bdevs": 2, 00:11:36.030 "num_base_bdevs_discovered": 2, 00:11:36.030 "num_base_bdevs_operational": 2, 00:11:36.030 "process": { 00:11:36.030 "type": "rebuild", 00:11:36.030 "target": "spare", 00:11:36.030 "progress": { 00:11:36.030 "blocks": 22528, 00:11:36.030 "percent": 35 00:11:36.030 } 00:11:36.030 }, 00:11:36.030 "base_bdevs_list": [ 00:11:36.030 { 00:11:36.030 "name": "spare", 00:11:36.030 "uuid": "01b0005d-400d-53f3-9990-eb6685bc4bd6", 00:11:36.030 "is_configured": true, 00:11:36.030 "data_offset": 2048, 00:11:36.030 "data_size": 63488 00:11:36.030 }, 00:11:36.030 { 00:11:36.030 "name": "BaseBdev2", 00:11:36.030 "uuid": "284114bb-d82a-54f3-a4eb-496e1224678c", 00:11:36.030 "is_configured": true, 00:11:36.030 "data_offset": 2048, 00:11:36.030 "data_size": 63488 00:11:36.030 } 00:11:36.030 ] 00:11:36.030 }' 00:11:36.030 15:16:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:36.030 15:16:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:36.030 15:16:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:36.030 15:16:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:36.030 15:16:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:36.985 15:16:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:36.985 15:16:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:36.985 15:16:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:36.985 15:16:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:36.985 15:16:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:36.985 15:16:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:36.985 15:16:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:36.985 15:16:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:36.985 15:16:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:36.985 15:16:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:36.985 15:16:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:36.985 15:16:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:36.985 "name": "raid_bdev1", 00:11:36.985 "uuid": "d9fe91ae-3d72-44a3-83a7-3a94d10441b0", 00:11:36.985 "strip_size_kb": 0, 00:11:36.985 "state": "online", 00:11:36.985 "raid_level": "raid1", 00:11:36.985 "superblock": true, 00:11:36.985 "num_base_bdevs": 2, 00:11:36.985 "num_base_bdevs_discovered": 2, 00:11:36.985 "num_base_bdevs_operational": 2, 00:11:36.985 "process": { 00:11:36.985 "type": "rebuild", 00:11:36.985 "target": "spare", 00:11:36.985 "progress": { 00:11:36.985 "blocks": 45056, 00:11:36.985 "percent": 70 00:11:36.985 } 00:11:36.985 }, 00:11:36.985 "base_bdevs_list": [ 00:11:36.985 { 00:11:36.985 "name": "spare", 00:11:36.985 "uuid": "01b0005d-400d-53f3-9990-eb6685bc4bd6", 00:11:36.985 "is_configured": true, 00:11:36.985 "data_offset": 2048, 00:11:36.985 "data_size": 63488 00:11:36.985 }, 00:11:36.985 { 00:11:36.985 "name": "BaseBdev2", 00:11:36.985 "uuid": "284114bb-d82a-54f3-a4eb-496e1224678c", 00:11:36.985 "is_configured": true, 00:11:36.985 "data_offset": 2048, 00:11:36.985 "data_size": 63488 00:11:36.985 } 00:11:36.985 ] 00:11:36.985 }' 00:11:37.245 15:16:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:37.245 15:16:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:37.245 15:16:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:37.245 15:16:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:37.245 15:16:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:37.814 [2024-11-27 15:16:05.876982] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:11:37.814 [2024-11-27 15:16:05.877190] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:11:37.814 [2024-11-27 15:16:05.877415] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:38.072 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:38.072 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:38.072 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:38.072 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:38.072 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:38.072 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:38.072 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:38.072 15:16:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.072 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:38.072 15:16:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:38.331 15:16:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.331 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:38.331 "name": "raid_bdev1", 00:11:38.331 "uuid": "d9fe91ae-3d72-44a3-83a7-3a94d10441b0", 00:11:38.331 "strip_size_kb": 0, 00:11:38.331 "state": "online", 00:11:38.331 "raid_level": "raid1", 00:11:38.331 "superblock": true, 00:11:38.331 "num_base_bdevs": 2, 00:11:38.331 "num_base_bdevs_discovered": 2, 00:11:38.331 "num_base_bdevs_operational": 2, 00:11:38.331 "base_bdevs_list": [ 00:11:38.331 { 00:11:38.331 "name": "spare", 00:11:38.331 "uuid": "01b0005d-400d-53f3-9990-eb6685bc4bd6", 00:11:38.331 "is_configured": true, 00:11:38.331 "data_offset": 2048, 00:11:38.331 "data_size": 63488 00:11:38.331 }, 00:11:38.331 { 00:11:38.331 "name": "BaseBdev2", 00:11:38.331 "uuid": "284114bb-d82a-54f3-a4eb-496e1224678c", 00:11:38.331 "is_configured": true, 00:11:38.331 "data_offset": 2048, 00:11:38.331 "data_size": 63488 00:11:38.331 } 00:11:38.331 ] 00:11:38.331 }' 00:11:38.331 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:38.331 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:11:38.331 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:38.331 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:11:38.331 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:11:38.331 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:38.331 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:38.331 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:38.331 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:38.331 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:38.331 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:38.331 15:16:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.331 15:16:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:38.331 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:38.331 15:16:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.331 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:38.331 "name": "raid_bdev1", 00:11:38.331 "uuid": "d9fe91ae-3d72-44a3-83a7-3a94d10441b0", 00:11:38.331 "strip_size_kb": 0, 00:11:38.331 "state": "online", 00:11:38.331 "raid_level": "raid1", 00:11:38.331 "superblock": true, 00:11:38.331 "num_base_bdevs": 2, 00:11:38.331 "num_base_bdevs_discovered": 2, 00:11:38.331 "num_base_bdevs_operational": 2, 00:11:38.331 "base_bdevs_list": [ 00:11:38.331 { 00:11:38.331 "name": "spare", 00:11:38.331 "uuid": "01b0005d-400d-53f3-9990-eb6685bc4bd6", 00:11:38.331 "is_configured": true, 00:11:38.331 "data_offset": 2048, 00:11:38.331 "data_size": 63488 00:11:38.331 }, 00:11:38.331 { 00:11:38.331 "name": "BaseBdev2", 00:11:38.331 "uuid": "284114bb-d82a-54f3-a4eb-496e1224678c", 00:11:38.331 "is_configured": true, 00:11:38.331 "data_offset": 2048, 00:11:38.331 "data_size": 63488 00:11:38.331 } 00:11:38.331 ] 00:11:38.331 }' 00:11:38.331 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:38.331 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:38.331 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:38.331 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:38.331 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:38.331 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:38.331 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:38.331 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:38.331 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:38.331 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:38.331 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:38.331 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:38.331 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:38.332 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:38.332 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:38.332 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:38.332 15:16:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.332 15:16:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:38.591 15:16:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.591 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:38.591 "name": "raid_bdev1", 00:11:38.591 "uuid": "d9fe91ae-3d72-44a3-83a7-3a94d10441b0", 00:11:38.591 "strip_size_kb": 0, 00:11:38.591 "state": "online", 00:11:38.591 "raid_level": "raid1", 00:11:38.591 "superblock": true, 00:11:38.591 "num_base_bdevs": 2, 00:11:38.591 "num_base_bdevs_discovered": 2, 00:11:38.591 "num_base_bdevs_operational": 2, 00:11:38.591 "base_bdevs_list": [ 00:11:38.591 { 00:11:38.591 "name": "spare", 00:11:38.591 "uuid": "01b0005d-400d-53f3-9990-eb6685bc4bd6", 00:11:38.591 "is_configured": true, 00:11:38.591 "data_offset": 2048, 00:11:38.591 "data_size": 63488 00:11:38.591 }, 00:11:38.591 { 00:11:38.591 "name": "BaseBdev2", 00:11:38.591 "uuid": "284114bb-d82a-54f3-a4eb-496e1224678c", 00:11:38.591 "is_configured": true, 00:11:38.591 "data_offset": 2048, 00:11:38.591 "data_size": 63488 00:11:38.591 } 00:11:38.591 ] 00:11:38.591 }' 00:11:38.591 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:38.591 15:16:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:38.851 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:38.851 15:16:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.851 15:16:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:38.851 [2024-11-27 15:16:06.828709] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:38.851 [2024-11-27 15:16:06.828866] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:38.851 [2024-11-27 15:16:06.829141] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:38.851 [2024-11-27 15:16:06.829299] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:38.851 [2024-11-27 15:16:06.829367] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:11:38.851 15:16:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.851 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:38.851 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:11:38.851 15:16:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.851 15:16:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:38.851 15:16:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.851 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:11:38.851 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:11:38.851 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:11:38.851 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:11:38.851 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:38.851 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:11:38.851 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:38.851 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:11:38.851 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:38.851 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:11:38.851 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:38.851 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:38.851 15:16:06 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:11:39.110 /dev/nbd0 00:11:39.110 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:39.110 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:39.110 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:11:39.111 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:11:39.111 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:39.111 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:39.111 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:11:39.111 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:11:39.111 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:39.111 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:39.111 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:39.111 1+0 records in 00:11:39.111 1+0 records out 00:11:39.111 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000331443 s, 12.4 MB/s 00:11:39.111 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:39.111 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:11:39.111 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:39.111 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:39.111 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:11:39.111 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:39.111 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:39.111 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:11:39.370 /dev/nbd1 00:11:39.370 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:11:39.370 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:11:39.370 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:11:39.370 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:11:39.370 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:39.370 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:39.370 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:11:39.370 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:11:39.370 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:39.370 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:39.370 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:39.370 1+0 records in 00:11:39.370 1+0 records out 00:11:39.370 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000433289 s, 9.5 MB/s 00:11:39.370 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:39.370 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:11:39.370 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:39.370 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:39.370 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:11:39.370 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:39.370 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:39.370 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:11:39.629 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:11:39.629 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:39.629 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:11:39.629 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:39.629 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:11:39.629 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:39.629 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:39.629 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:39.629 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:39.629 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:39.629 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:39.629 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:39.629 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:39.629 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:11:39.629 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:11:39.629 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:39.629 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:11:39.888 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:11:39.888 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:11:39.888 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:11:39.888 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:39.888 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:39.888 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:11:39.888 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:11:39.888 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:11:39.889 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:11:39.889 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:11:39.889 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:39.889 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:39.889 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:39.889 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:39.889 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:39.889 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:39.889 [2024-11-27 15:16:07.957294] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:39.889 [2024-11-27 15:16:07.957377] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:39.889 [2024-11-27 15:16:07.957405] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:11:39.889 [2024-11-27 15:16:07.957422] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:39.889 [2024-11-27 15:16:07.960183] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:39.889 [2024-11-27 15:16:07.960279] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:39.889 [2024-11-27 15:16:07.960416] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:11:39.889 [2024-11-27 15:16:07.960508] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:39.889 [2024-11-27 15:16:07.960714] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:39.889 spare 00:11:39.889 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:39.889 15:16:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:11:39.889 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:39.889 15:16:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:40.148 [2024-11-27 15:16:08.060706] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006600 00:11:40.148 [2024-11-27 15:16:08.060794] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:40.148 [2024-11-27 15:16:08.061230] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1940 00:11:40.148 [2024-11-27 15:16:08.061485] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006600 00:11:40.148 [2024-11-27 15:16:08.061540] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006600 00:11:40.148 [2024-11-27 15:16:08.061780] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:40.148 15:16:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:40.148 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:40.148 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:40.148 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:40.148 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:40.148 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:40.148 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:40.148 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:40.148 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:40.148 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:40.148 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:40.148 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:40.148 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:40.148 15:16:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:40.148 15:16:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:40.148 15:16:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:40.148 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:40.148 "name": "raid_bdev1", 00:11:40.148 "uuid": "d9fe91ae-3d72-44a3-83a7-3a94d10441b0", 00:11:40.148 "strip_size_kb": 0, 00:11:40.148 "state": "online", 00:11:40.148 "raid_level": "raid1", 00:11:40.148 "superblock": true, 00:11:40.148 "num_base_bdevs": 2, 00:11:40.148 "num_base_bdevs_discovered": 2, 00:11:40.148 "num_base_bdevs_operational": 2, 00:11:40.148 "base_bdevs_list": [ 00:11:40.148 { 00:11:40.148 "name": "spare", 00:11:40.148 "uuid": "01b0005d-400d-53f3-9990-eb6685bc4bd6", 00:11:40.148 "is_configured": true, 00:11:40.148 "data_offset": 2048, 00:11:40.148 "data_size": 63488 00:11:40.148 }, 00:11:40.148 { 00:11:40.148 "name": "BaseBdev2", 00:11:40.148 "uuid": "284114bb-d82a-54f3-a4eb-496e1224678c", 00:11:40.148 "is_configured": true, 00:11:40.148 "data_offset": 2048, 00:11:40.148 "data_size": 63488 00:11:40.148 } 00:11:40.148 ] 00:11:40.148 }' 00:11:40.148 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:40.149 15:16:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:40.408 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:40.408 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:40.408 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:40.408 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:40.408 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:40.408 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:40.408 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:40.408 15:16:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:40.408 15:16:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:40.667 "name": "raid_bdev1", 00:11:40.667 "uuid": "d9fe91ae-3d72-44a3-83a7-3a94d10441b0", 00:11:40.667 "strip_size_kb": 0, 00:11:40.667 "state": "online", 00:11:40.667 "raid_level": "raid1", 00:11:40.667 "superblock": true, 00:11:40.667 "num_base_bdevs": 2, 00:11:40.667 "num_base_bdevs_discovered": 2, 00:11:40.667 "num_base_bdevs_operational": 2, 00:11:40.667 "base_bdevs_list": [ 00:11:40.667 { 00:11:40.667 "name": "spare", 00:11:40.667 "uuid": "01b0005d-400d-53f3-9990-eb6685bc4bd6", 00:11:40.667 "is_configured": true, 00:11:40.667 "data_offset": 2048, 00:11:40.667 "data_size": 63488 00:11:40.667 }, 00:11:40.667 { 00:11:40.667 "name": "BaseBdev2", 00:11:40.667 "uuid": "284114bb-d82a-54f3-a4eb-496e1224678c", 00:11:40.667 "is_configured": true, 00:11:40.667 "data_offset": 2048, 00:11:40.667 "data_size": 63488 00:11:40.667 } 00:11:40.667 ] 00:11:40.667 }' 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:40.667 [2024-11-27 15:16:08.704691] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:40.667 "name": "raid_bdev1", 00:11:40.667 "uuid": "d9fe91ae-3d72-44a3-83a7-3a94d10441b0", 00:11:40.667 "strip_size_kb": 0, 00:11:40.667 "state": "online", 00:11:40.667 "raid_level": "raid1", 00:11:40.667 "superblock": true, 00:11:40.667 "num_base_bdevs": 2, 00:11:40.667 "num_base_bdevs_discovered": 1, 00:11:40.667 "num_base_bdevs_operational": 1, 00:11:40.667 "base_bdevs_list": [ 00:11:40.667 { 00:11:40.667 "name": null, 00:11:40.667 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:40.667 "is_configured": false, 00:11:40.667 "data_offset": 0, 00:11:40.667 "data_size": 63488 00:11:40.667 }, 00:11:40.667 { 00:11:40.667 "name": "BaseBdev2", 00:11:40.667 "uuid": "284114bb-d82a-54f3-a4eb-496e1224678c", 00:11:40.667 "is_configured": true, 00:11:40.667 "data_offset": 2048, 00:11:40.667 "data_size": 63488 00:11:40.667 } 00:11:40.667 ] 00:11:40.667 }' 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:40.667 15:16:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:41.235 15:16:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:41.235 15:16:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.235 15:16:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:41.235 [2024-11-27 15:16:09.176000] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:41.235 [2024-11-27 15:16:09.176326] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:11:41.235 [2024-11-27 15:16:09.176353] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:11:41.235 [2024-11-27 15:16:09.176412] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:41.235 [2024-11-27 15:16:09.185323] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1a10 00:11:41.235 15:16:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.235 15:16:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:11:41.235 [2024-11-27 15:16:09.187706] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:42.173 15:16:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:42.173 15:16:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:42.173 15:16:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:42.173 15:16:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:42.173 15:16:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:42.173 15:16:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:42.173 15:16:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:42.173 15:16:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:42.173 15:16:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:42.173 15:16:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:42.173 15:16:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:42.173 "name": "raid_bdev1", 00:11:42.173 "uuid": "d9fe91ae-3d72-44a3-83a7-3a94d10441b0", 00:11:42.173 "strip_size_kb": 0, 00:11:42.173 "state": "online", 00:11:42.173 "raid_level": "raid1", 00:11:42.173 "superblock": true, 00:11:42.173 "num_base_bdevs": 2, 00:11:42.173 "num_base_bdevs_discovered": 2, 00:11:42.173 "num_base_bdevs_operational": 2, 00:11:42.173 "process": { 00:11:42.173 "type": "rebuild", 00:11:42.173 "target": "spare", 00:11:42.173 "progress": { 00:11:42.173 "blocks": 20480, 00:11:42.173 "percent": 32 00:11:42.173 } 00:11:42.173 }, 00:11:42.173 "base_bdevs_list": [ 00:11:42.173 { 00:11:42.173 "name": "spare", 00:11:42.173 "uuid": "01b0005d-400d-53f3-9990-eb6685bc4bd6", 00:11:42.173 "is_configured": true, 00:11:42.173 "data_offset": 2048, 00:11:42.173 "data_size": 63488 00:11:42.173 }, 00:11:42.173 { 00:11:42.173 "name": "BaseBdev2", 00:11:42.173 "uuid": "284114bb-d82a-54f3-a4eb-496e1224678c", 00:11:42.173 "is_configured": true, 00:11:42.173 "data_offset": 2048, 00:11:42.173 "data_size": 63488 00:11:42.173 } 00:11:42.173 ] 00:11:42.173 }' 00:11:42.173 15:16:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:42.173 15:16:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:42.173 15:16:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:42.431 15:16:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:42.431 15:16:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:11:42.431 15:16:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:42.431 15:16:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:42.431 [2024-11-27 15:16:10.324388] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:42.431 [2024-11-27 15:16:10.397035] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:42.431 [2024-11-27 15:16:10.397136] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:42.431 [2024-11-27 15:16:10.397163] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:42.431 [2024-11-27 15:16:10.397174] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:42.431 15:16:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:42.431 15:16:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:42.431 15:16:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:42.431 15:16:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:42.431 15:16:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:42.431 15:16:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:42.431 15:16:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:42.431 15:16:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:42.431 15:16:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:42.431 15:16:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:42.431 15:16:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:42.431 15:16:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:42.431 15:16:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:42.431 15:16:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:42.431 15:16:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:42.431 15:16:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:42.431 15:16:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:42.431 "name": "raid_bdev1", 00:11:42.431 "uuid": "d9fe91ae-3d72-44a3-83a7-3a94d10441b0", 00:11:42.431 "strip_size_kb": 0, 00:11:42.431 "state": "online", 00:11:42.431 "raid_level": "raid1", 00:11:42.431 "superblock": true, 00:11:42.431 "num_base_bdevs": 2, 00:11:42.431 "num_base_bdevs_discovered": 1, 00:11:42.431 "num_base_bdevs_operational": 1, 00:11:42.431 "base_bdevs_list": [ 00:11:42.431 { 00:11:42.431 "name": null, 00:11:42.431 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:42.431 "is_configured": false, 00:11:42.432 "data_offset": 0, 00:11:42.432 "data_size": 63488 00:11:42.432 }, 00:11:42.432 { 00:11:42.432 "name": "BaseBdev2", 00:11:42.432 "uuid": "284114bb-d82a-54f3-a4eb-496e1224678c", 00:11:42.432 "is_configured": true, 00:11:42.432 "data_offset": 2048, 00:11:42.432 "data_size": 63488 00:11:42.432 } 00:11:42.432 ] 00:11:42.432 }' 00:11:42.432 15:16:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:42.432 15:16:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:43.001 15:16:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:43.001 15:16:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:43.001 15:16:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:43.001 [2024-11-27 15:16:10.837262] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:43.001 [2024-11-27 15:16:10.837369] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:43.001 [2024-11-27 15:16:10.837408] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:11:43.001 [2024-11-27 15:16:10.837423] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:43.001 [2024-11-27 15:16:10.838116] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:43.001 [2024-11-27 15:16:10.838156] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:43.001 [2024-11-27 15:16:10.838295] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:11:43.001 [2024-11-27 15:16:10.838314] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:11:43.001 [2024-11-27 15:16:10.838335] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:11:43.001 [2024-11-27 15:16:10.838378] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:43.001 [2024-11-27 15:16:10.847472] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1ae0 00:11:43.001 spare 00:11:43.001 15:16:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:43.001 15:16:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:11:43.001 [2024-11-27 15:16:10.850130] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:43.941 15:16:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:43.941 15:16:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:43.941 15:16:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:43.941 15:16:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:43.941 15:16:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:43.941 15:16:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:43.941 15:16:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:43.941 15:16:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:43.941 15:16:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:43.941 15:16:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:43.941 15:16:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:43.941 "name": "raid_bdev1", 00:11:43.941 "uuid": "d9fe91ae-3d72-44a3-83a7-3a94d10441b0", 00:11:43.941 "strip_size_kb": 0, 00:11:43.941 "state": "online", 00:11:43.941 "raid_level": "raid1", 00:11:43.941 "superblock": true, 00:11:43.941 "num_base_bdevs": 2, 00:11:43.941 "num_base_bdevs_discovered": 2, 00:11:43.941 "num_base_bdevs_operational": 2, 00:11:43.941 "process": { 00:11:43.941 "type": "rebuild", 00:11:43.941 "target": "spare", 00:11:43.941 "progress": { 00:11:43.941 "blocks": 20480, 00:11:43.941 "percent": 32 00:11:43.941 } 00:11:43.941 }, 00:11:43.941 "base_bdevs_list": [ 00:11:43.941 { 00:11:43.941 "name": "spare", 00:11:43.941 "uuid": "01b0005d-400d-53f3-9990-eb6685bc4bd6", 00:11:43.941 "is_configured": true, 00:11:43.941 "data_offset": 2048, 00:11:43.941 "data_size": 63488 00:11:43.941 }, 00:11:43.941 { 00:11:43.941 "name": "BaseBdev2", 00:11:43.941 "uuid": "284114bb-d82a-54f3-a4eb-496e1224678c", 00:11:43.941 "is_configured": true, 00:11:43.941 "data_offset": 2048, 00:11:43.941 "data_size": 63488 00:11:43.941 } 00:11:43.941 ] 00:11:43.941 }' 00:11:43.941 15:16:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:43.941 15:16:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:43.941 15:16:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:43.941 15:16:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:43.941 15:16:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:11:43.941 15:16:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:43.941 15:16:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:43.941 [2024-11-27 15:16:12.011265] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:44.200 [2024-11-27 15:16:12.059991] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:44.200 [2024-11-27 15:16:12.060181] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:44.200 [2024-11-27 15:16:12.060236] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:44.200 [2024-11-27 15:16:12.060273] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:44.200 15:16:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:44.200 15:16:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:44.200 15:16:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:44.200 15:16:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:44.200 15:16:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:44.200 15:16:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:44.200 15:16:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:44.200 15:16:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:44.200 15:16:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:44.201 15:16:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:44.201 15:16:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:44.201 15:16:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:44.201 15:16:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:44.201 15:16:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:44.201 15:16:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:44.201 15:16:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:44.201 15:16:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:44.201 "name": "raid_bdev1", 00:11:44.201 "uuid": "d9fe91ae-3d72-44a3-83a7-3a94d10441b0", 00:11:44.201 "strip_size_kb": 0, 00:11:44.201 "state": "online", 00:11:44.201 "raid_level": "raid1", 00:11:44.201 "superblock": true, 00:11:44.201 "num_base_bdevs": 2, 00:11:44.201 "num_base_bdevs_discovered": 1, 00:11:44.201 "num_base_bdevs_operational": 1, 00:11:44.201 "base_bdevs_list": [ 00:11:44.201 { 00:11:44.201 "name": null, 00:11:44.201 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:44.201 "is_configured": false, 00:11:44.201 "data_offset": 0, 00:11:44.201 "data_size": 63488 00:11:44.201 }, 00:11:44.201 { 00:11:44.201 "name": "BaseBdev2", 00:11:44.201 "uuid": "284114bb-d82a-54f3-a4eb-496e1224678c", 00:11:44.201 "is_configured": true, 00:11:44.201 "data_offset": 2048, 00:11:44.201 "data_size": 63488 00:11:44.201 } 00:11:44.201 ] 00:11:44.201 }' 00:11:44.201 15:16:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:44.201 15:16:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:44.460 15:16:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:44.460 15:16:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:44.460 15:16:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:44.460 15:16:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:44.460 15:16:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:44.460 15:16:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:44.460 15:16:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:44.460 15:16:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:44.460 15:16:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:44.460 15:16:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:44.725 15:16:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:44.725 "name": "raid_bdev1", 00:11:44.725 "uuid": "d9fe91ae-3d72-44a3-83a7-3a94d10441b0", 00:11:44.725 "strip_size_kb": 0, 00:11:44.725 "state": "online", 00:11:44.725 "raid_level": "raid1", 00:11:44.725 "superblock": true, 00:11:44.725 "num_base_bdevs": 2, 00:11:44.725 "num_base_bdevs_discovered": 1, 00:11:44.725 "num_base_bdevs_operational": 1, 00:11:44.725 "base_bdevs_list": [ 00:11:44.725 { 00:11:44.725 "name": null, 00:11:44.725 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:44.725 "is_configured": false, 00:11:44.725 "data_offset": 0, 00:11:44.725 "data_size": 63488 00:11:44.725 }, 00:11:44.725 { 00:11:44.725 "name": "BaseBdev2", 00:11:44.726 "uuid": "284114bb-d82a-54f3-a4eb-496e1224678c", 00:11:44.726 "is_configured": true, 00:11:44.726 "data_offset": 2048, 00:11:44.726 "data_size": 63488 00:11:44.726 } 00:11:44.726 ] 00:11:44.726 }' 00:11:44.726 15:16:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:44.726 15:16:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:44.726 15:16:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:44.726 15:16:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:44.726 15:16:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:11:44.726 15:16:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:44.726 15:16:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:44.726 15:16:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:44.726 15:16:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:44.726 15:16:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:44.726 15:16:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:44.726 [2024-11-27 15:16:12.676224] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:44.726 [2024-11-27 15:16:12.676328] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:44.726 [2024-11-27 15:16:12.676360] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:11:44.726 [2024-11-27 15:16:12.676377] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:44.726 [2024-11-27 15:16:12.676988] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:44.726 [2024-11-27 15:16:12.677109] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:44.726 [2024-11-27 15:16:12.677238] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:11:44.726 [2024-11-27 15:16:12.677271] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:11:44.726 [2024-11-27 15:16:12.677284] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:11:44.726 [2024-11-27 15:16:12.677309] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:11:44.726 BaseBdev1 00:11:44.726 15:16:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:44.726 15:16:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:11:45.673 15:16:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:45.673 15:16:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:45.673 15:16:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:45.673 15:16:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:45.673 15:16:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:45.673 15:16:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:45.673 15:16:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:45.673 15:16:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:45.673 15:16:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:45.673 15:16:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:45.673 15:16:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:45.673 15:16:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:45.673 15:16:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.673 15:16:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:45.673 15:16:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.673 15:16:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:45.673 "name": "raid_bdev1", 00:11:45.673 "uuid": "d9fe91ae-3d72-44a3-83a7-3a94d10441b0", 00:11:45.673 "strip_size_kb": 0, 00:11:45.673 "state": "online", 00:11:45.673 "raid_level": "raid1", 00:11:45.673 "superblock": true, 00:11:45.673 "num_base_bdevs": 2, 00:11:45.674 "num_base_bdevs_discovered": 1, 00:11:45.674 "num_base_bdevs_operational": 1, 00:11:45.674 "base_bdevs_list": [ 00:11:45.674 { 00:11:45.674 "name": null, 00:11:45.674 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:45.674 "is_configured": false, 00:11:45.674 "data_offset": 0, 00:11:45.674 "data_size": 63488 00:11:45.674 }, 00:11:45.674 { 00:11:45.674 "name": "BaseBdev2", 00:11:45.674 "uuid": "284114bb-d82a-54f3-a4eb-496e1224678c", 00:11:45.674 "is_configured": true, 00:11:45.674 "data_offset": 2048, 00:11:45.674 "data_size": 63488 00:11:45.674 } 00:11:45.674 ] 00:11:45.674 }' 00:11:45.674 15:16:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:45.674 15:16:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:46.241 15:16:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:46.241 15:16:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:46.241 15:16:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:46.241 15:16:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:46.241 15:16:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:46.241 15:16:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:46.241 15:16:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:46.241 15:16:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:46.241 15:16:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:46.241 15:16:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:46.241 15:16:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:46.241 "name": "raid_bdev1", 00:11:46.241 "uuid": "d9fe91ae-3d72-44a3-83a7-3a94d10441b0", 00:11:46.241 "strip_size_kb": 0, 00:11:46.241 "state": "online", 00:11:46.241 "raid_level": "raid1", 00:11:46.241 "superblock": true, 00:11:46.241 "num_base_bdevs": 2, 00:11:46.241 "num_base_bdevs_discovered": 1, 00:11:46.241 "num_base_bdevs_operational": 1, 00:11:46.241 "base_bdevs_list": [ 00:11:46.241 { 00:11:46.241 "name": null, 00:11:46.241 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:46.241 "is_configured": false, 00:11:46.241 "data_offset": 0, 00:11:46.241 "data_size": 63488 00:11:46.241 }, 00:11:46.241 { 00:11:46.241 "name": "BaseBdev2", 00:11:46.241 "uuid": "284114bb-d82a-54f3-a4eb-496e1224678c", 00:11:46.241 "is_configured": true, 00:11:46.241 "data_offset": 2048, 00:11:46.241 "data_size": 63488 00:11:46.241 } 00:11:46.241 ] 00:11:46.241 }' 00:11:46.241 15:16:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:46.241 15:16:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:46.241 15:16:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:46.241 15:16:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:46.241 15:16:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:11:46.241 15:16:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@652 -- # local es=0 00:11:46.241 15:16:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:11:46.241 15:16:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:11:46.241 15:16:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:11:46.241 15:16:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:11:46.241 15:16:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:11:46.241 15:16:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:11:46.241 15:16:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:46.241 15:16:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:46.241 [2024-11-27 15:16:14.263520] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:46.241 [2024-11-27 15:16:14.263814] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:11:46.241 [2024-11-27 15:16:14.263896] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:11:46.241 request: 00:11:46.241 { 00:11:46.241 "base_bdev": "BaseBdev1", 00:11:46.241 "raid_bdev": "raid_bdev1", 00:11:46.241 "method": "bdev_raid_add_base_bdev", 00:11:46.241 "req_id": 1 00:11:46.241 } 00:11:46.241 Got JSON-RPC error response 00:11:46.241 response: 00:11:46.241 { 00:11:46.241 "code": -22, 00:11:46.241 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:11:46.241 } 00:11:46.241 15:16:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:11:46.241 15:16:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@655 -- # es=1 00:11:46.241 15:16:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:11:46.241 15:16:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:11:46.241 15:16:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:11:46.241 15:16:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:11:47.176 15:16:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:47.176 15:16:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:47.177 15:16:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:47.177 15:16:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:47.177 15:16:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:47.177 15:16:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:47.177 15:16:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:47.177 15:16:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:47.177 15:16:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:47.177 15:16:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:47.435 15:16:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:47.435 15:16:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:47.435 15:16:15 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:47.435 15:16:15 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:47.435 15:16:15 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:47.435 15:16:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:47.435 "name": "raid_bdev1", 00:11:47.435 "uuid": "d9fe91ae-3d72-44a3-83a7-3a94d10441b0", 00:11:47.435 "strip_size_kb": 0, 00:11:47.435 "state": "online", 00:11:47.435 "raid_level": "raid1", 00:11:47.435 "superblock": true, 00:11:47.435 "num_base_bdevs": 2, 00:11:47.435 "num_base_bdevs_discovered": 1, 00:11:47.435 "num_base_bdevs_operational": 1, 00:11:47.435 "base_bdevs_list": [ 00:11:47.435 { 00:11:47.435 "name": null, 00:11:47.435 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:47.435 "is_configured": false, 00:11:47.435 "data_offset": 0, 00:11:47.435 "data_size": 63488 00:11:47.435 }, 00:11:47.435 { 00:11:47.435 "name": "BaseBdev2", 00:11:47.435 "uuid": "284114bb-d82a-54f3-a4eb-496e1224678c", 00:11:47.435 "is_configured": true, 00:11:47.435 "data_offset": 2048, 00:11:47.435 "data_size": 63488 00:11:47.435 } 00:11:47.435 ] 00:11:47.435 }' 00:11:47.435 15:16:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:47.435 15:16:15 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:47.694 15:16:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:47.694 15:16:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:47.694 15:16:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:47.694 15:16:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:47.694 15:16:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:47.694 15:16:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:47.694 15:16:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:47.694 15:16:15 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:47.694 15:16:15 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:47.694 15:16:15 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:47.694 15:16:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:47.694 "name": "raid_bdev1", 00:11:47.694 "uuid": "d9fe91ae-3d72-44a3-83a7-3a94d10441b0", 00:11:47.694 "strip_size_kb": 0, 00:11:47.694 "state": "online", 00:11:47.694 "raid_level": "raid1", 00:11:47.694 "superblock": true, 00:11:47.694 "num_base_bdevs": 2, 00:11:47.694 "num_base_bdevs_discovered": 1, 00:11:47.694 "num_base_bdevs_operational": 1, 00:11:47.694 "base_bdevs_list": [ 00:11:47.694 { 00:11:47.694 "name": null, 00:11:47.694 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:47.694 "is_configured": false, 00:11:47.694 "data_offset": 0, 00:11:47.694 "data_size": 63488 00:11:47.694 }, 00:11:47.694 { 00:11:47.694 "name": "BaseBdev2", 00:11:47.694 "uuid": "284114bb-d82a-54f3-a4eb-496e1224678c", 00:11:47.694 "is_configured": true, 00:11:47.694 "data_offset": 2048, 00:11:47.694 "data_size": 63488 00:11:47.694 } 00:11:47.694 ] 00:11:47.694 }' 00:11:47.694 15:16:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:47.953 15:16:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:47.953 15:16:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:47.953 15:16:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:47.953 15:16:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 86572 00:11:47.953 15:16:15 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@954 -- # '[' -z 86572 ']' 00:11:47.953 15:16:15 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@958 -- # kill -0 86572 00:11:47.953 15:16:15 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@959 -- # uname 00:11:47.953 15:16:15 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:47.953 15:16:15 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 86572 00:11:47.953 killing process with pid 86572 00:11:47.953 Received shutdown signal, test time was about 60.000000 seconds 00:11:47.953 00:11:47.953 Latency(us) 00:11:47.953 [2024-11-27T15:16:16.060Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:47.953 [2024-11-27T15:16:16.060Z] =================================================================================================================== 00:11:47.953 [2024-11-27T15:16:16.060Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:11:47.953 15:16:15 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:47.953 15:16:15 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:47.953 15:16:15 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 86572' 00:11:47.953 15:16:15 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@973 -- # kill 86572 00:11:47.953 [2024-11-27 15:16:15.919027] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:47.953 15:16:15 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@978 -- # wait 86572 00:11:47.953 [2024-11-27 15:16:15.919229] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:47.953 [2024-11-27 15:16:15.919315] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:47.953 [2024-11-27 15:16:15.919329] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state offline 00:11:47.953 [2024-11-27 15:16:15.979030] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:11:48.522 00:11:48.522 real 0m22.411s 00:11:48.522 user 0m27.021s 00:11:48.522 sys 0m4.001s 00:11:48.522 ************************************ 00:11:48.522 END TEST raid_rebuild_test_sb 00:11:48.522 ************************************ 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:48.522 15:16:16 bdev_raid -- bdev/bdev_raid.sh@980 -- # run_test raid_rebuild_test_io raid_rebuild_test raid1 2 false true true 00:11:48.522 15:16:16 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:11:48.522 15:16:16 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:48.522 15:16:16 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:48.522 ************************************ 00:11:48.522 START TEST raid_rebuild_test_io 00:11:48.522 ************************************ 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 false true true 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@597 -- # raid_pid=87294 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 87294 00:11:48.522 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@835 -- # '[' -z 87294 ']' 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:48.522 15:16:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:48.522 [2024-11-27 15:16:16.515208] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:11:48.522 I/O size of 3145728 is greater than zero copy threshold (65536). 00:11:48.522 Zero copy mechanism will not be used. 00:11:48.522 [2024-11-27 15:16:16.515497] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87294 ] 00:11:48.780 [2024-11-27 15:16:16.693927] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:48.780 [2024-11-27 15:16:16.721571] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:48.780 [2024-11-27 15:16:16.765131] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:48.780 [2024-11-27 15:16:16.765213] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:49.348 15:16:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:49.348 15:16:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # return 0 00:11:49.348 15:16:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:49.348 15:16:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:49.348 15:16:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.348 15:16:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:49.348 BaseBdev1_malloc 00:11:49.348 15:16:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:49.348 15:16:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:49.348 15:16:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.348 15:16:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:49.348 [2024-11-27 15:16:17.401354] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:49.348 [2024-11-27 15:16:17.401473] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:49.348 [2024-11-27 15:16:17.401518] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:11:49.348 [2024-11-27 15:16:17.401531] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:49.348 [2024-11-27 15:16:17.403659] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:49.348 [2024-11-27 15:16:17.403696] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:49.348 BaseBdev1 00:11:49.348 15:16:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:49.348 15:16:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:49.348 15:16:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:49.348 15:16:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.348 15:16:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:49.348 BaseBdev2_malloc 00:11:49.348 15:16:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:49.348 15:16:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:11:49.348 15:16:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.348 15:16:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:49.348 [2024-11-27 15:16:17.429925] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:11:49.348 [2024-11-27 15:16:17.429976] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:49.348 [2024-11-27 15:16:17.430011] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:11:49.348 [2024-11-27 15:16:17.430020] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:49.348 [2024-11-27 15:16:17.432068] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:49.348 [2024-11-27 15:16:17.432158] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:49.348 BaseBdev2 00:11:49.348 15:16:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:49.348 15:16:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:11:49.348 15:16:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.348 15:16:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:49.607 spare_malloc 00:11:49.607 15:16:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:49.607 15:16:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:11:49.607 15:16:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.607 15:16:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:49.607 spare_delay 00:11:49.607 15:16:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:49.607 15:16:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:49.607 15:16:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.607 15:16:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:49.607 [2024-11-27 15:16:17.470497] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:49.607 [2024-11-27 15:16:17.470602] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:49.607 [2024-11-27 15:16:17.470643] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:11:49.607 [2024-11-27 15:16:17.470652] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:49.607 [2024-11-27 15:16:17.472904] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:49.607 [2024-11-27 15:16:17.472945] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:49.607 spare 00:11:49.607 15:16:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:49.607 15:16:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:11:49.607 15:16:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.607 15:16:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:49.607 [2024-11-27 15:16:17.482504] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:49.607 [2024-11-27 15:16:17.484281] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:49.607 [2024-11-27 15:16:17.484373] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:11:49.607 [2024-11-27 15:16:17.484384] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:11:49.607 [2024-11-27 15:16:17.484635] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:11:49.607 [2024-11-27 15:16:17.484778] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:11:49.607 [2024-11-27 15:16:17.484792] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:11:49.607 [2024-11-27 15:16:17.484893] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:49.607 15:16:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:49.607 15:16:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:49.607 15:16:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:49.607 15:16:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:49.607 15:16:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:49.607 15:16:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:49.607 15:16:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:49.607 15:16:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:49.607 15:16:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:49.607 15:16:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:49.607 15:16:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:49.607 15:16:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:49.607 15:16:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:49.607 15:16:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.607 15:16:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:49.607 15:16:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:49.607 15:16:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:49.607 "name": "raid_bdev1", 00:11:49.607 "uuid": "d1bfdf69-bc74-45b9-aa5d-552e33f6a3b5", 00:11:49.607 "strip_size_kb": 0, 00:11:49.607 "state": "online", 00:11:49.607 "raid_level": "raid1", 00:11:49.607 "superblock": false, 00:11:49.607 "num_base_bdevs": 2, 00:11:49.607 "num_base_bdevs_discovered": 2, 00:11:49.607 "num_base_bdevs_operational": 2, 00:11:49.607 "base_bdevs_list": [ 00:11:49.607 { 00:11:49.607 "name": "BaseBdev1", 00:11:49.607 "uuid": "b95c2f58-a738-5098-b811-a92dad8523de", 00:11:49.607 "is_configured": true, 00:11:49.607 "data_offset": 0, 00:11:49.607 "data_size": 65536 00:11:49.607 }, 00:11:49.607 { 00:11:49.607 "name": "BaseBdev2", 00:11:49.607 "uuid": "91af75e7-1e10-5606-bd66-2eb3df1ae0c9", 00:11:49.607 "is_configured": true, 00:11:49.607 "data_offset": 0, 00:11:49.607 "data_size": 65536 00:11:49.607 } 00:11:49.607 ] 00:11:49.607 }' 00:11:49.607 15:16:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:49.607 15:16:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:50.181 15:16:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:11:50.181 15:16:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:50.181 15:16:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:50.181 15:16:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:50.181 [2024-11-27 15:16:17.981989] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:50.181 15:16:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:50.181 15:16:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:11:50.181 15:16:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:50.181 15:16:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:50.181 15:16:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:50.181 15:16:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:11:50.181 15:16:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:50.181 15:16:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:11:50.181 15:16:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:11:50.181 15:16:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:11:50.181 15:16:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:11:50.181 15:16:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:50.181 15:16:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:50.181 [2024-11-27 15:16:18.081480] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:50.181 15:16:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:50.181 15:16:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:50.181 15:16:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:50.181 15:16:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:50.181 15:16:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:50.181 15:16:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:50.181 15:16:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:50.181 15:16:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:50.181 15:16:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:50.181 15:16:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:50.181 15:16:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:50.181 15:16:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:50.181 15:16:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:50.181 15:16:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:50.181 15:16:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:50.181 15:16:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:50.181 15:16:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:50.181 "name": "raid_bdev1", 00:11:50.181 "uuid": "d1bfdf69-bc74-45b9-aa5d-552e33f6a3b5", 00:11:50.181 "strip_size_kb": 0, 00:11:50.181 "state": "online", 00:11:50.181 "raid_level": "raid1", 00:11:50.181 "superblock": false, 00:11:50.181 "num_base_bdevs": 2, 00:11:50.181 "num_base_bdevs_discovered": 1, 00:11:50.181 "num_base_bdevs_operational": 1, 00:11:50.181 "base_bdevs_list": [ 00:11:50.181 { 00:11:50.181 "name": null, 00:11:50.181 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:50.181 "is_configured": false, 00:11:50.181 "data_offset": 0, 00:11:50.181 "data_size": 65536 00:11:50.181 }, 00:11:50.181 { 00:11:50.181 "name": "BaseBdev2", 00:11:50.181 "uuid": "91af75e7-1e10-5606-bd66-2eb3df1ae0c9", 00:11:50.181 "is_configured": true, 00:11:50.181 "data_offset": 0, 00:11:50.181 "data_size": 65536 00:11:50.181 } 00:11:50.181 ] 00:11:50.181 }' 00:11:50.181 15:16:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:50.181 15:16:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:50.181 [2024-11-27 15:16:18.182173] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:11:50.181 I/O size of 3145728 is greater than zero copy threshold (65536). 00:11:50.181 Zero copy mechanism will not be used. 00:11:50.181 Running I/O for 60 seconds... 00:11:50.443 15:16:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:50.443 15:16:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:50.443 15:16:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:50.443 [2024-11-27 15:16:18.543793] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:50.703 15:16:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:50.703 15:16:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:11:50.703 [2024-11-27 15:16:18.593574] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:11:50.703 [2024-11-27 15:16:18.595486] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:50.703 [2024-11-27 15:16:18.709240] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:50.703 [2024-11-27 15:16:18.709848] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:50.963 [2024-11-27 15:16:18.923922] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:50.963 [2024-11-27 15:16:18.924305] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:51.222 182.00 IOPS, 546.00 MiB/s [2024-11-27T15:16:19.329Z] [2024-11-27 15:16:19.280518] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:51.222 [2024-11-27 15:16:19.280989] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:51.481 15:16:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:51.481 15:16:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:51.481 15:16:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:51.481 15:16:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:51.481 15:16:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:51.481 15:16:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:51.481 15:16:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:51.481 15:16:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:51.481 15:16:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:51.740 15:16:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:51.740 [2024-11-27 15:16:19.618149] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:11:51.740 [2024-11-27 15:16:19.618385] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:11:51.740 15:16:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:51.740 "name": "raid_bdev1", 00:11:51.740 "uuid": "d1bfdf69-bc74-45b9-aa5d-552e33f6a3b5", 00:11:51.740 "strip_size_kb": 0, 00:11:51.740 "state": "online", 00:11:51.740 "raid_level": "raid1", 00:11:51.740 "superblock": false, 00:11:51.740 "num_base_bdevs": 2, 00:11:51.740 "num_base_bdevs_discovered": 2, 00:11:51.740 "num_base_bdevs_operational": 2, 00:11:51.740 "process": { 00:11:51.740 "type": "rebuild", 00:11:51.740 "target": "spare", 00:11:51.740 "progress": { 00:11:51.740 "blocks": 14336, 00:11:51.740 "percent": 21 00:11:51.740 } 00:11:51.740 }, 00:11:51.740 "base_bdevs_list": [ 00:11:51.740 { 00:11:51.740 "name": "spare", 00:11:51.740 "uuid": "03356e84-e1e0-5588-8d37-cd511de5ccf9", 00:11:51.740 "is_configured": true, 00:11:51.740 "data_offset": 0, 00:11:51.740 "data_size": 65536 00:11:51.740 }, 00:11:51.740 { 00:11:51.740 "name": "BaseBdev2", 00:11:51.740 "uuid": "91af75e7-1e10-5606-bd66-2eb3df1ae0c9", 00:11:51.740 "is_configured": true, 00:11:51.740 "data_offset": 0, 00:11:51.740 "data_size": 65536 00:11:51.740 } 00:11:51.740 ] 00:11:51.740 }' 00:11:51.740 15:16:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:51.740 15:16:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:51.740 15:16:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:51.740 15:16:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:51.740 15:16:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:51.740 15:16:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:51.740 15:16:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:51.740 [2024-11-27 15:16:19.718117] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:51.740 [2024-11-27 15:16:19.833648] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:51.740 [2024-11-27 15:16:19.841677] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:51.740 [2024-11-27 15:16:19.841799] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:51.740 [2024-11-27 15:16:19.841818] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:52.001 [2024-11-27 15:16:19.864960] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000005ee0 00:11:52.001 15:16:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.001 15:16:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:52.001 15:16:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:52.001 15:16:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:52.001 15:16:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:52.001 15:16:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:52.001 15:16:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:52.001 15:16:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:52.001 15:16:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:52.001 15:16:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:52.001 15:16:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:52.001 15:16:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:52.001 15:16:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:52.001 15:16:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.001 15:16:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:52.001 15:16:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.001 15:16:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:52.001 "name": "raid_bdev1", 00:11:52.001 "uuid": "d1bfdf69-bc74-45b9-aa5d-552e33f6a3b5", 00:11:52.001 "strip_size_kb": 0, 00:11:52.001 "state": "online", 00:11:52.001 "raid_level": "raid1", 00:11:52.001 "superblock": false, 00:11:52.001 "num_base_bdevs": 2, 00:11:52.001 "num_base_bdevs_discovered": 1, 00:11:52.001 "num_base_bdevs_operational": 1, 00:11:52.001 "base_bdevs_list": [ 00:11:52.001 { 00:11:52.001 "name": null, 00:11:52.001 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:52.001 "is_configured": false, 00:11:52.001 "data_offset": 0, 00:11:52.001 "data_size": 65536 00:11:52.001 }, 00:11:52.001 { 00:11:52.001 "name": "BaseBdev2", 00:11:52.001 "uuid": "91af75e7-1e10-5606-bd66-2eb3df1ae0c9", 00:11:52.001 "is_configured": true, 00:11:52.001 "data_offset": 0, 00:11:52.001 "data_size": 65536 00:11:52.001 } 00:11:52.001 ] 00:11:52.001 }' 00:11:52.001 15:16:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:52.001 15:16:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:52.262 154.50 IOPS, 463.50 MiB/s [2024-11-27T15:16:20.369Z] 15:16:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:52.262 15:16:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:52.262 15:16:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:52.262 15:16:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:52.262 15:16:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:52.262 15:16:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:52.262 15:16:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.262 15:16:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:52.262 15:16:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:52.262 15:16:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.522 15:16:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:52.522 "name": "raid_bdev1", 00:11:52.522 "uuid": "d1bfdf69-bc74-45b9-aa5d-552e33f6a3b5", 00:11:52.522 "strip_size_kb": 0, 00:11:52.522 "state": "online", 00:11:52.522 "raid_level": "raid1", 00:11:52.522 "superblock": false, 00:11:52.522 "num_base_bdevs": 2, 00:11:52.522 "num_base_bdevs_discovered": 1, 00:11:52.522 "num_base_bdevs_operational": 1, 00:11:52.522 "base_bdevs_list": [ 00:11:52.522 { 00:11:52.522 "name": null, 00:11:52.522 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:52.522 "is_configured": false, 00:11:52.522 "data_offset": 0, 00:11:52.522 "data_size": 65536 00:11:52.522 }, 00:11:52.522 { 00:11:52.522 "name": "BaseBdev2", 00:11:52.522 "uuid": "91af75e7-1e10-5606-bd66-2eb3df1ae0c9", 00:11:52.522 "is_configured": true, 00:11:52.522 "data_offset": 0, 00:11:52.522 "data_size": 65536 00:11:52.522 } 00:11:52.522 ] 00:11:52.522 }' 00:11:52.522 15:16:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:52.522 15:16:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:52.522 15:16:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:52.522 15:16:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:52.522 15:16:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:52.522 15:16:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.522 15:16:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:52.522 [2024-11-27 15:16:20.477320] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:52.522 15:16:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.522 15:16:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:11:52.522 [2024-11-27 15:16:20.526699] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:11:52.522 [2024-11-27 15:16:20.528760] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:52.782 [2024-11-27 15:16:20.635572] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:52.782 [2024-11-27 15:16:20.636020] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:52.782 [2024-11-27 15:16:20.849702] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:52.782 [2024-11-27 15:16:20.850090] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:53.352 [2024-11-27 15:16:21.185012] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:11:53.352 154.33 IOPS, 463.00 MiB/s [2024-11-27T15:16:21.459Z] [2024-11-27 15:16:21.401029] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:53.614 "name": "raid_bdev1", 00:11:53.614 "uuid": "d1bfdf69-bc74-45b9-aa5d-552e33f6a3b5", 00:11:53.614 "strip_size_kb": 0, 00:11:53.614 "state": "online", 00:11:53.614 "raid_level": "raid1", 00:11:53.614 "superblock": false, 00:11:53.614 "num_base_bdevs": 2, 00:11:53.614 "num_base_bdevs_discovered": 2, 00:11:53.614 "num_base_bdevs_operational": 2, 00:11:53.614 "process": { 00:11:53.614 "type": "rebuild", 00:11:53.614 "target": "spare", 00:11:53.614 "progress": { 00:11:53.614 "blocks": 10240, 00:11:53.614 "percent": 15 00:11:53.614 } 00:11:53.614 }, 00:11:53.614 "base_bdevs_list": [ 00:11:53.614 { 00:11:53.614 "name": "spare", 00:11:53.614 "uuid": "03356e84-e1e0-5588-8d37-cd511de5ccf9", 00:11:53.614 "is_configured": true, 00:11:53.614 "data_offset": 0, 00:11:53.614 "data_size": 65536 00:11:53.614 }, 00:11:53.614 { 00:11:53.614 "name": "BaseBdev2", 00:11:53.614 "uuid": "91af75e7-1e10-5606-bd66-2eb3df1ae0c9", 00:11:53.614 "is_configured": true, 00:11:53.614 "data_offset": 0, 00:11:53.614 "data_size": 65536 00:11:53.614 } 00:11:53.614 ] 00:11:53.614 }' 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@706 -- # local timeout=322 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:53.614 "name": "raid_bdev1", 00:11:53.614 "uuid": "d1bfdf69-bc74-45b9-aa5d-552e33f6a3b5", 00:11:53.614 "strip_size_kb": 0, 00:11:53.614 "state": "online", 00:11:53.614 "raid_level": "raid1", 00:11:53.614 "superblock": false, 00:11:53.614 "num_base_bdevs": 2, 00:11:53.614 "num_base_bdevs_discovered": 2, 00:11:53.614 "num_base_bdevs_operational": 2, 00:11:53.614 "process": { 00:11:53.614 "type": "rebuild", 00:11:53.614 "target": "spare", 00:11:53.614 "progress": { 00:11:53.614 "blocks": 12288, 00:11:53.614 "percent": 18 00:11:53.614 } 00:11:53.614 }, 00:11:53.614 "base_bdevs_list": [ 00:11:53.614 { 00:11:53.614 "name": "spare", 00:11:53.614 "uuid": "03356e84-e1e0-5588-8d37-cd511de5ccf9", 00:11:53.614 "is_configured": true, 00:11:53.614 "data_offset": 0, 00:11:53.614 "data_size": 65536 00:11:53.614 }, 00:11:53.614 { 00:11:53.614 "name": "BaseBdev2", 00:11:53.614 "uuid": "91af75e7-1e10-5606-bd66-2eb3df1ae0c9", 00:11:53.614 "is_configured": true, 00:11:53.614 "data_offset": 0, 00:11:53.614 "data_size": 65536 00:11:53.614 } 00:11:53.614 ] 00:11:53.614 }' 00:11:53.614 15:16:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:53.950 15:16:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:53.950 15:16:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:53.950 [2024-11-27 15:16:21.736162] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:11:53.950 15:16:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:53.950 15:16:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:53.950 [2024-11-27 15:16:21.963489] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:11:53.950 [2024-11-27 15:16:21.963763] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:11:54.210 130.75 IOPS, 392.25 MiB/s [2024-11-27T15:16:22.317Z] [2024-11-27 15:16:22.185679] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:11:54.210 [2024-11-27 15:16:22.186185] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:11:54.469 [2024-11-27 15:16:22.405712] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:11:54.469 [2024-11-27 15:16:22.406000] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:11:54.729 [2024-11-27 15:16:22.726868] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:11:54.729 15:16:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:54.729 15:16:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:54.729 15:16:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:54.729 15:16:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:54.729 15:16:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:54.729 15:16:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:54.729 15:16:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:54.729 15:16:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:54.729 15:16:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:54.729 15:16:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:54.729 15:16:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:54.988 15:16:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:54.988 "name": "raid_bdev1", 00:11:54.988 "uuid": "d1bfdf69-bc74-45b9-aa5d-552e33f6a3b5", 00:11:54.988 "strip_size_kb": 0, 00:11:54.988 "state": "online", 00:11:54.988 "raid_level": "raid1", 00:11:54.988 "superblock": false, 00:11:54.988 "num_base_bdevs": 2, 00:11:54.988 "num_base_bdevs_discovered": 2, 00:11:54.988 "num_base_bdevs_operational": 2, 00:11:54.988 "process": { 00:11:54.988 "type": "rebuild", 00:11:54.988 "target": "spare", 00:11:54.988 "progress": { 00:11:54.988 "blocks": 26624, 00:11:54.988 "percent": 40 00:11:54.988 } 00:11:54.988 }, 00:11:54.988 "base_bdevs_list": [ 00:11:54.988 { 00:11:54.988 "name": "spare", 00:11:54.988 "uuid": "03356e84-e1e0-5588-8d37-cd511de5ccf9", 00:11:54.988 "is_configured": true, 00:11:54.988 "data_offset": 0, 00:11:54.988 "data_size": 65536 00:11:54.988 }, 00:11:54.988 { 00:11:54.988 "name": "BaseBdev2", 00:11:54.988 "uuid": "91af75e7-1e10-5606-bd66-2eb3df1ae0c9", 00:11:54.988 "is_configured": true, 00:11:54.988 "data_offset": 0, 00:11:54.988 "data_size": 65536 00:11:54.988 } 00:11:54.988 ] 00:11:54.988 }' 00:11:54.988 15:16:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:54.988 15:16:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:54.988 15:16:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:54.988 15:16:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:54.988 15:16:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:54.988 [2024-11-27 15:16:22.959539] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 28672 offset_begin: 24576 offset_end: 30720 00:11:55.507 114.40 IOPS, 343.20 MiB/s [2024-11-27T15:16:23.614Z] [2024-11-27 15:16:23.393398] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:11:55.507 [2024-11-27 15:16:23.393653] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:11:55.767 [2024-11-27 15:16:23.846123] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 40960 offset_begin: 36864 offset_end: 43008 00:11:56.028 15:16:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:56.028 15:16:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:56.028 15:16:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:56.028 15:16:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:56.028 15:16:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:56.028 15:16:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:56.028 15:16:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:56.028 15:16:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:56.028 15:16:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:56.028 15:16:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:56.028 15:16:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:56.028 15:16:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:56.028 "name": "raid_bdev1", 00:11:56.028 "uuid": "d1bfdf69-bc74-45b9-aa5d-552e33f6a3b5", 00:11:56.028 "strip_size_kb": 0, 00:11:56.028 "state": "online", 00:11:56.028 "raid_level": "raid1", 00:11:56.028 "superblock": false, 00:11:56.028 "num_base_bdevs": 2, 00:11:56.028 "num_base_bdevs_discovered": 2, 00:11:56.028 "num_base_bdevs_operational": 2, 00:11:56.028 "process": { 00:11:56.028 "type": "rebuild", 00:11:56.028 "target": "spare", 00:11:56.028 "progress": { 00:11:56.028 "blocks": 40960, 00:11:56.028 "percent": 62 00:11:56.028 } 00:11:56.028 }, 00:11:56.028 "base_bdevs_list": [ 00:11:56.028 { 00:11:56.028 "name": "spare", 00:11:56.028 "uuid": "03356e84-e1e0-5588-8d37-cd511de5ccf9", 00:11:56.028 "is_configured": true, 00:11:56.028 "data_offset": 0, 00:11:56.028 "data_size": 65536 00:11:56.028 }, 00:11:56.028 { 00:11:56.028 "name": "BaseBdev2", 00:11:56.028 "uuid": "91af75e7-1e10-5606-bd66-2eb3df1ae0c9", 00:11:56.028 "is_configured": true, 00:11:56.028 "data_offset": 0, 00:11:56.028 "data_size": 65536 00:11:56.028 } 00:11:56.028 ] 00:11:56.028 }' 00:11:56.028 15:16:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:56.028 15:16:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:56.028 15:16:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:56.028 15:16:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:56.028 15:16:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:56.288 [2024-11-27 15:16:24.176346] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 45056 offset_begin: 43008 offset_end: 49152 00:11:56.548 102.83 IOPS, 308.50 MiB/s [2024-11-27T15:16:24.655Z] [2024-11-27 15:16:24.617026] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 51200 offset_begin: 49152 offset_end: 55296 00:11:57.120 15:16:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:57.120 15:16:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:57.120 15:16:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:57.120 15:16:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:57.120 15:16:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:57.120 15:16:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:57.120 15:16:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:57.120 15:16:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:57.120 15:16:25 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:57.120 15:16:25 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:57.120 15:16:25 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:57.120 15:16:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:57.120 "name": "raid_bdev1", 00:11:57.120 "uuid": "d1bfdf69-bc74-45b9-aa5d-552e33f6a3b5", 00:11:57.120 "strip_size_kb": 0, 00:11:57.120 "state": "online", 00:11:57.120 "raid_level": "raid1", 00:11:57.120 "superblock": false, 00:11:57.120 "num_base_bdevs": 2, 00:11:57.120 "num_base_bdevs_discovered": 2, 00:11:57.120 "num_base_bdevs_operational": 2, 00:11:57.120 "process": { 00:11:57.120 "type": "rebuild", 00:11:57.120 "target": "spare", 00:11:57.120 "progress": { 00:11:57.120 "blocks": 57344, 00:11:57.120 "percent": 87 00:11:57.120 } 00:11:57.120 }, 00:11:57.120 "base_bdevs_list": [ 00:11:57.120 { 00:11:57.120 "name": "spare", 00:11:57.120 "uuid": "03356e84-e1e0-5588-8d37-cd511de5ccf9", 00:11:57.120 "is_configured": true, 00:11:57.120 "data_offset": 0, 00:11:57.120 "data_size": 65536 00:11:57.120 }, 00:11:57.120 { 00:11:57.120 "name": "BaseBdev2", 00:11:57.120 "uuid": "91af75e7-1e10-5606-bd66-2eb3df1ae0c9", 00:11:57.120 "is_configured": true, 00:11:57.120 "data_offset": 0, 00:11:57.120 "data_size": 65536 00:11:57.120 } 00:11:57.120 ] 00:11:57.120 }' 00:11:57.120 15:16:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:57.120 15:16:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:57.120 15:16:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:57.120 92.71 IOPS, 278.14 MiB/s [2024-11-27T15:16:25.227Z] 15:16:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:57.120 15:16:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:57.380 [2024-11-27 15:16:25.477028] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:11:57.640 [2024-11-27 15:16:25.582151] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:11:57.640 [2024-11-27 15:16:25.584812] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:58.210 85.12 IOPS, 255.38 MiB/s [2024-11-27T15:16:26.317Z] 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:58.210 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:58.210 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:58.210 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:58.210 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:58.210 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:58.210 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:58.210 15:16:26 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.210 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:58.210 15:16:26 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:58.210 15:16:26 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.210 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:58.210 "name": "raid_bdev1", 00:11:58.210 "uuid": "d1bfdf69-bc74-45b9-aa5d-552e33f6a3b5", 00:11:58.210 "strip_size_kb": 0, 00:11:58.210 "state": "online", 00:11:58.210 "raid_level": "raid1", 00:11:58.210 "superblock": false, 00:11:58.210 "num_base_bdevs": 2, 00:11:58.210 "num_base_bdevs_discovered": 2, 00:11:58.210 "num_base_bdevs_operational": 2, 00:11:58.210 "base_bdevs_list": [ 00:11:58.210 { 00:11:58.210 "name": "spare", 00:11:58.210 "uuid": "03356e84-e1e0-5588-8d37-cd511de5ccf9", 00:11:58.210 "is_configured": true, 00:11:58.210 "data_offset": 0, 00:11:58.210 "data_size": 65536 00:11:58.210 }, 00:11:58.210 { 00:11:58.210 "name": "BaseBdev2", 00:11:58.210 "uuid": "91af75e7-1e10-5606-bd66-2eb3df1ae0c9", 00:11:58.210 "is_configured": true, 00:11:58.210 "data_offset": 0, 00:11:58.210 "data_size": 65536 00:11:58.210 } 00:11:58.210 ] 00:11:58.210 }' 00:11:58.210 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:58.470 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:11:58.470 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:58.470 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:11:58.470 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@709 -- # break 00:11:58.470 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:58.470 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:58.470 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:58.470 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:58.470 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:58.470 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:58.470 15:16:26 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.470 15:16:26 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:58.470 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:58.470 15:16:26 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.470 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:58.470 "name": "raid_bdev1", 00:11:58.470 "uuid": "d1bfdf69-bc74-45b9-aa5d-552e33f6a3b5", 00:11:58.470 "strip_size_kb": 0, 00:11:58.470 "state": "online", 00:11:58.470 "raid_level": "raid1", 00:11:58.470 "superblock": false, 00:11:58.470 "num_base_bdevs": 2, 00:11:58.470 "num_base_bdevs_discovered": 2, 00:11:58.470 "num_base_bdevs_operational": 2, 00:11:58.470 "base_bdevs_list": [ 00:11:58.470 { 00:11:58.470 "name": "spare", 00:11:58.470 "uuid": "03356e84-e1e0-5588-8d37-cd511de5ccf9", 00:11:58.470 "is_configured": true, 00:11:58.470 "data_offset": 0, 00:11:58.470 "data_size": 65536 00:11:58.470 }, 00:11:58.470 { 00:11:58.470 "name": "BaseBdev2", 00:11:58.470 "uuid": "91af75e7-1e10-5606-bd66-2eb3df1ae0c9", 00:11:58.470 "is_configured": true, 00:11:58.470 "data_offset": 0, 00:11:58.470 "data_size": 65536 00:11:58.470 } 00:11:58.470 ] 00:11:58.470 }' 00:11:58.470 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:58.470 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:58.470 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:58.470 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:58.470 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:58.470 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:58.470 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:58.470 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:58.471 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:58.471 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:58.471 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:58.471 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:58.471 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:58.471 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:58.471 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:58.471 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:58.471 15:16:26 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.471 15:16:26 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:58.471 15:16:26 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.471 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:58.471 "name": "raid_bdev1", 00:11:58.471 "uuid": "d1bfdf69-bc74-45b9-aa5d-552e33f6a3b5", 00:11:58.471 "strip_size_kb": 0, 00:11:58.471 "state": "online", 00:11:58.471 "raid_level": "raid1", 00:11:58.471 "superblock": false, 00:11:58.471 "num_base_bdevs": 2, 00:11:58.471 "num_base_bdevs_discovered": 2, 00:11:58.471 "num_base_bdevs_operational": 2, 00:11:58.471 "base_bdevs_list": [ 00:11:58.471 { 00:11:58.471 "name": "spare", 00:11:58.471 "uuid": "03356e84-e1e0-5588-8d37-cd511de5ccf9", 00:11:58.471 "is_configured": true, 00:11:58.471 "data_offset": 0, 00:11:58.471 "data_size": 65536 00:11:58.471 }, 00:11:58.471 { 00:11:58.471 "name": "BaseBdev2", 00:11:58.471 "uuid": "91af75e7-1e10-5606-bd66-2eb3df1ae0c9", 00:11:58.471 "is_configured": true, 00:11:58.471 "data_offset": 0, 00:11:58.471 "data_size": 65536 00:11:58.471 } 00:11:58.471 ] 00:11:58.471 }' 00:11:58.471 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:58.471 15:16:26 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:59.039 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:59.039 15:16:26 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:59.039 15:16:26 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:59.039 [2024-11-27 15:16:26.890363] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:59.039 [2024-11-27 15:16:26.890394] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:59.039 00:11:59.039 Latency(us) 00:11:59.039 [2024-11-27T15:16:27.146Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:59.039 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:11:59.039 raid_bdev1 : 8.81 81.04 243.11 0.00 0.00 16913.95 289.76 114473.36 00:11:59.039 [2024-11-27T15:16:27.146Z] =================================================================================================================== 00:11:59.039 [2024-11-27T15:16:27.146Z] Total : 81.04 243.11 0.00 0.00 16913.95 289.76 114473.36 00:11:59.039 [2024-11-27 15:16:26.981662] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:59.039 [2024-11-27 15:16:26.981722] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:59.039 [2024-11-27 15:16:26.981793] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:59.039 [2024-11-27 15:16:26.981805] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:11:59.039 { 00:11:59.039 "results": [ 00:11:59.039 { 00:11:59.039 "job": "raid_bdev1", 00:11:59.040 "core_mask": "0x1", 00:11:59.040 "workload": "randrw", 00:11:59.040 "percentage": 50, 00:11:59.040 "status": "finished", 00:11:59.040 "queue_depth": 2, 00:11:59.040 "io_size": 3145728, 00:11:59.040 "runtime": 8.81075, 00:11:59.040 "iops": 81.0373691229464, 00:11:59.040 "mibps": 243.1121073688392, 00:11:59.040 "io_failed": 0, 00:11:59.040 "io_timeout": 0, 00:11:59.040 "avg_latency_us": 16913.94909055325, 00:11:59.040 "min_latency_us": 289.7606986899563, 00:11:59.040 "max_latency_us": 114473.36244541485 00:11:59.040 } 00:11:59.040 ], 00:11:59.040 "core_count": 1 00:11:59.040 } 00:11:59.040 15:16:26 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:59.040 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:59.040 15:16:26 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:59.040 15:16:26 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:59.040 15:16:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # jq length 00:11:59.040 15:16:26 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:59.040 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:11:59.040 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:11:59.040 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:11:59.040 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:11:59.040 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:59.040 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:11:59.040 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:59.040 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:11:59.040 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:59.040 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:11:59.040 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:59.040 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:59.040 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:11:59.300 /dev/nbd0 00:11:59.300 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:59.300 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:59.300 15:16:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:11:59.300 15:16:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:11:59.300 15:16:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:59.300 15:16:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:59.301 15:16:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:11:59.301 15:16:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:11:59.301 15:16:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:59.301 15:16:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:59.301 15:16:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:59.301 1+0 records in 00:11:59.301 1+0 records out 00:11:59.301 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000316868 s, 12.9 MB/s 00:11:59.301 15:16:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:59.301 15:16:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:11:59.301 15:16:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:59.301 15:16:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:59.301 15:16:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:11:59.301 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:59.301 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:59.301 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:11:59.301 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev2 ']' 00:11:59.301 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev2 /dev/nbd1 00:11:59.301 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:59.301 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev2') 00:11:59.301 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:59.301 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:11:59.301 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:59.301 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:11:59.301 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:59.301 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:59.301 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev2 /dev/nbd1 00:11:59.561 /dev/nbd1 00:11:59.561 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:11:59.561 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:11:59.561 15:16:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:11:59.561 15:16:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:11:59.561 15:16:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:59.561 15:16:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:59.561 15:16:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:11:59.561 15:16:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:11:59.561 15:16:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:59.561 15:16:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:59.561 15:16:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:59.561 1+0 records in 00:11:59.561 1+0 records out 00:11:59.561 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000526236 s, 7.8 MB/s 00:11:59.562 15:16:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:59.562 15:16:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:11:59.562 15:16:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:59.562 15:16:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:59.562 15:16:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:11:59.562 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:59.562 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:59.562 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:11:59.562 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:11:59.562 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:59.562 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:11:59.562 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:59.562 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:11:59.562 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:59.562 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:11:59.822 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:11:59.822 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:11:59.822 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:11:59.822 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:59.822 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:59.822 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:11:59.822 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:11:59.822 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:11:59.822 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:11:59.822 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:59.822 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:11:59.822 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:59.822 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:11:59.822 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:59.822 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:00.082 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:00.082 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:00.082 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:00.082 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:00.082 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:00.082 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:00.082 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:12:00.082 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:00.082 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:12:00.082 15:16:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@784 -- # killprocess 87294 00:12:00.082 15:16:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@954 -- # '[' -z 87294 ']' 00:12:00.082 15:16:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@958 -- # kill -0 87294 00:12:00.082 15:16:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@959 -- # uname 00:12:00.082 15:16:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:00.082 15:16:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 87294 00:12:00.082 killing process with pid 87294 00:12:00.082 Received shutdown signal, test time was about 9.868553 seconds 00:12:00.082 00:12:00.082 Latency(us) 00:12:00.082 [2024-11-27T15:16:28.189Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:00.082 [2024-11-27T15:16:28.189Z] =================================================================================================================== 00:12:00.082 [2024-11-27T15:16:28.189Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:12:00.082 15:16:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:00.082 15:16:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:00.082 15:16:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@972 -- # echo 'killing process with pid 87294' 00:12:00.082 15:16:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@973 -- # kill 87294 00:12:00.082 [2024-11-27 15:16:28.033924] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:00.082 15:16:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@978 -- # wait 87294 00:12:00.082 [2024-11-27 15:16:28.080844] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:00.342 15:16:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@786 -- # return 0 00:12:00.342 00:12:00.342 real 0m12.016s 00:12:00.342 user 0m15.281s 00:12:00.342 sys 0m1.449s 00:12:00.342 15:16:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:00.342 15:16:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:00.342 ************************************ 00:12:00.342 END TEST raid_rebuild_test_io 00:12:00.342 ************************************ 00:12:00.602 15:16:28 bdev_raid -- bdev/bdev_raid.sh@981 -- # run_test raid_rebuild_test_sb_io raid_rebuild_test raid1 2 true true true 00:12:00.602 15:16:28 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:12:00.602 15:16:28 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:00.602 15:16:28 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:00.602 ************************************ 00:12:00.602 START TEST raid_rebuild_test_sb_io 00:12:00.602 ************************************ 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true true true 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@597 -- # raid_pid=87680 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 87680 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@835 -- # '[' -z 87680 ']' 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:00.602 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:00.602 15:16:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:00.602 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:00.602 Zero copy mechanism will not be used. 00:12:00.602 [2024-11-27 15:16:28.579718] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:12:00.602 [2024-11-27 15:16:28.579872] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87680 ] 00:12:00.861 [2024-11-27 15:16:28.748245] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:00.861 [2024-11-27 15:16:28.786883] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:00.861 [2024-11-27 15:16:28.862819] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:00.861 [2024-11-27 15:16:28.862864] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:01.429 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:01.429 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # return 0 00:12:01.429 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:01.429 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:01.429 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.429 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:01.429 BaseBdev1_malloc 00:12:01.429 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.429 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:01.429 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.429 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:01.429 [2024-11-27 15:16:29.424516] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:01.429 [2024-11-27 15:16:29.424604] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:01.429 [2024-11-27 15:16:29.424641] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:01.429 [2024-11-27 15:16:29.424665] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:01.429 [2024-11-27 15:16:29.427153] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:01.429 [2024-11-27 15:16:29.427193] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:01.429 BaseBdev1 00:12:01.429 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.429 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:01.429 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:01.429 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.429 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:01.429 BaseBdev2_malloc 00:12:01.429 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.429 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:12:01.429 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.429 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:01.429 [2024-11-27 15:16:29.459317] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:12:01.429 [2024-11-27 15:16:29.459382] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:01.429 [2024-11-27 15:16:29.459408] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:12:01.429 [2024-11-27 15:16:29.459419] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:01.429 [2024-11-27 15:16:29.461835] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:01.429 [2024-11-27 15:16:29.461875] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:01.429 BaseBdev2 00:12:01.429 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.429 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:12:01.429 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.429 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:01.429 spare_malloc 00:12:01.429 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.429 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:12:01.429 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.429 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:01.429 spare_delay 00:12:01.429 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.430 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:01.430 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.430 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:01.430 [2024-11-27 15:16:29.506062] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:01.430 [2024-11-27 15:16:29.506128] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:01.430 [2024-11-27 15:16:29.506154] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:12:01.430 [2024-11-27 15:16:29.506164] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:01.430 [2024-11-27 15:16:29.508569] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:01.430 [2024-11-27 15:16:29.508610] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:01.430 spare 00:12:01.430 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.430 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:12:01.430 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.430 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:01.430 [2024-11-27 15:16:29.518091] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:01.430 [2024-11-27 15:16:29.520194] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:01.430 [2024-11-27 15:16:29.520373] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:12:01.430 [2024-11-27 15:16:29.520399] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:01.430 [2024-11-27 15:16:29.520720] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:12:01.430 [2024-11-27 15:16:29.520916] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:12:01.430 [2024-11-27 15:16:29.520939] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:12:01.430 [2024-11-27 15:16:29.521078] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:01.430 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.430 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:01.430 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:01.430 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:01.430 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:01.430 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:01.430 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:01.430 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:01.430 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:01.430 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:01.430 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:01.430 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:01.430 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:01.430 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.430 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:01.689 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.689 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:01.689 "name": "raid_bdev1", 00:12:01.689 "uuid": "ff205561-c290-495f-8446-929b386e84dd", 00:12:01.689 "strip_size_kb": 0, 00:12:01.689 "state": "online", 00:12:01.689 "raid_level": "raid1", 00:12:01.689 "superblock": true, 00:12:01.689 "num_base_bdevs": 2, 00:12:01.689 "num_base_bdevs_discovered": 2, 00:12:01.689 "num_base_bdevs_operational": 2, 00:12:01.689 "base_bdevs_list": [ 00:12:01.689 { 00:12:01.689 "name": "BaseBdev1", 00:12:01.689 "uuid": "6165c030-5d9a-56d1-8ea7-a8a4cf3058f5", 00:12:01.690 "is_configured": true, 00:12:01.690 "data_offset": 2048, 00:12:01.690 "data_size": 63488 00:12:01.690 }, 00:12:01.690 { 00:12:01.690 "name": "BaseBdev2", 00:12:01.690 "uuid": "477c21a6-2fad-53aa-8334-14a66e35f763", 00:12:01.690 "is_configured": true, 00:12:01.690 "data_offset": 2048, 00:12:01.690 "data_size": 63488 00:12:01.690 } 00:12:01.690 ] 00:12:01.690 }' 00:12:01.690 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:01.690 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:01.950 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:01.950 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.950 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:12:01.950 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:01.950 [2024-11-27 15:16:29.905745] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:01.950 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.950 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:12:01.950 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:01.950 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:12:01.950 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.950 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:01.950 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.950 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:12:01.950 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:12:01.950 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:12:01.950 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:12:01.950 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.950 15:16:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:01.950 [2024-11-27 15:16:30.001288] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:01.950 15:16:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.950 15:16:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:01.950 15:16:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:01.950 15:16:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:01.950 15:16:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:01.950 15:16:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:01.950 15:16:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:01.950 15:16:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:01.950 15:16:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:01.950 15:16:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:01.950 15:16:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:01.950 15:16:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:01.950 15:16:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:01.950 15:16:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.950 15:16:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:01.950 15:16:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.950 15:16:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:01.950 "name": "raid_bdev1", 00:12:01.950 "uuid": "ff205561-c290-495f-8446-929b386e84dd", 00:12:01.950 "strip_size_kb": 0, 00:12:01.950 "state": "online", 00:12:01.950 "raid_level": "raid1", 00:12:01.950 "superblock": true, 00:12:01.950 "num_base_bdevs": 2, 00:12:01.950 "num_base_bdevs_discovered": 1, 00:12:01.950 "num_base_bdevs_operational": 1, 00:12:01.950 "base_bdevs_list": [ 00:12:01.950 { 00:12:01.950 "name": null, 00:12:01.950 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:01.950 "is_configured": false, 00:12:01.950 "data_offset": 0, 00:12:01.950 "data_size": 63488 00:12:01.950 }, 00:12:01.950 { 00:12:01.950 "name": "BaseBdev2", 00:12:01.950 "uuid": "477c21a6-2fad-53aa-8334-14a66e35f763", 00:12:01.950 "is_configured": true, 00:12:01.950 "data_offset": 2048, 00:12:01.950 "data_size": 63488 00:12:01.950 } 00:12:01.950 ] 00:12:01.950 }' 00:12:01.950 15:16:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:01.950 15:16:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:02.210 [2024-11-27 15:16:30.104284] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:12:02.210 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:02.210 Zero copy mechanism will not be used. 00:12:02.210 Running I/O for 60 seconds... 00:12:02.469 15:16:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:02.469 15:16:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.469 15:16:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:02.469 [2024-11-27 15:16:30.457299] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:02.469 15:16:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.469 15:16:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:12:02.469 [2024-11-27 15:16:30.506779] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:12:02.469 [2024-11-27 15:16:30.509130] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:02.729 [2024-11-27 15:16:30.629639] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:02.729 [2024-11-27 15:16:30.630389] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:02.988 [2024-11-27 15:16:30.862022] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:02.988 [2024-11-27 15:16:30.862433] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:03.334 191.00 IOPS, 573.00 MiB/s [2024-11-27T15:16:31.441Z] [2024-11-27 15:16:31.221179] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:03.622 [2024-11-27 15:16:31.443812] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:03.622 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:03.622 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:03.622 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:03.622 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:03.622 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:03.622 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:03.622 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:03.622 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.622 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:03.622 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.622 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:03.622 "name": "raid_bdev1", 00:12:03.622 "uuid": "ff205561-c290-495f-8446-929b386e84dd", 00:12:03.622 "strip_size_kb": 0, 00:12:03.622 "state": "online", 00:12:03.622 "raid_level": "raid1", 00:12:03.622 "superblock": true, 00:12:03.622 "num_base_bdevs": 2, 00:12:03.622 "num_base_bdevs_discovered": 2, 00:12:03.622 "num_base_bdevs_operational": 2, 00:12:03.622 "process": { 00:12:03.622 "type": "rebuild", 00:12:03.622 "target": "spare", 00:12:03.622 "progress": { 00:12:03.622 "blocks": 10240, 00:12:03.622 "percent": 16 00:12:03.622 } 00:12:03.622 }, 00:12:03.622 "base_bdevs_list": [ 00:12:03.622 { 00:12:03.622 "name": "spare", 00:12:03.622 "uuid": "5ff3b68f-1f8b-5b9b-b9e3-6a342f601efb", 00:12:03.622 "is_configured": true, 00:12:03.622 "data_offset": 2048, 00:12:03.622 "data_size": 63488 00:12:03.622 }, 00:12:03.622 { 00:12:03.622 "name": "BaseBdev2", 00:12:03.622 "uuid": "477c21a6-2fad-53aa-8334-14a66e35f763", 00:12:03.622 "is_configured": true, 00:12:03.622 "data_offset": 2048, 00:12:03.622 "data_size": 63488 00:12:03.622 } 00:12:03.622 ] 00:12:03.622 }' 00:12:03.622 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:03.622 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:03.622 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:03.622 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:03.622 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:03.622 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.622 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:03.622 [2024-11-27 15:16:31.646531] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:03.882 [2024-11-27 15:16:31.768211] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:03.882 [2024-11-27 15:16:31.771428] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:03.882 [2024-11-27 15:16:31.771494] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:03.882 [2024-11-27 15:16:31.771525] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:03.882 [2024-11-27 15:16:31.787832] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000005ee0 00:12:03.882 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.882 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:03.882 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:03.882 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:03.882 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:03.882 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:03.882 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:03.882 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:03.882 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:03.882 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:03.882 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:03.882 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:03.882 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.882 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:03.882 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:03.882 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.882 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:03.882 "name": "raid_bdev1", 00:12:03.882 "uuid": "ff205561-c290-495f-8446-929b386e84dd", 00:12:03.882 "strip_size_kb": 0, 00:12:03.882 "state": "online", 00:12:03.882 "raid_level": "raid1", 00:12:03.882 "superblock": true, 00:12:03.882 "num_base_bdevs": 2, 00:12:03.882 "num_base_bdevs_discovered": 1, 00:12:03.882 "num_base_bdevs_operational": 1, 00:12:03.882 "base_bdevs_list": [ 00:12:03.882 { 00:12:03.882 "name": null, 00:12:03.882 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:03.882 "is_configured": false, 00:12:03.882 "data_offset": 0, 00:12:03.882 "data_size": 63488 00:12:03.882 }, 00:12:03.882 { 00:12:03.882 "name": "BaseBdev2", 00:12:03.882 "uuid": "477c21a6-2fad-53aa-8334-14a66e35f763", 00:12:03.882 "is_configured": true, 00:12:03.882 "data_offset": 2048, 00:12:03.882 "data_size": 63488 00:12:03.882 } 00:12:03.882 ] 00:12:03.882 }' 00:12:03.882 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:03.882 15:16:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:04.141 163.50 IOPS, 490.50 MiB/s [2024-11-27T15:16:32.248Z] 15:16:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:04.141 15:16:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:04.141 15:16:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:04.141 15:16:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:04.141 15:16:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:04.141 15:16:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:04.141 15:16:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:04.141 15:16:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:04.141 15:16:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:04.400 15:16:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:04.400 15:16:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:04.400 "name": "raid_bdev1", 00:12:04.400 "uuid": "ff205561-c290-495f-8446-929b386e84dd", 00:12:04.400 "strip_size_kb": 0, 00:12:04.400 "state": "online", 00:12:04.400 "raid_level": "raid1", 00:12:04.400 "superblock": true, 00:12:04.400 "num_base_bdevs": 2, 00:12:04.400 "num_base_bdevs_discovered": 1, 00:12:04.400 "num_base_bdevs_operational": 1, 00:12:04.400 "base_bdevs_list": [ 00:12:04.400 { 00:12:04.400 "name": null, 00:12:04.400 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:04.400 "is_configured": false, 00:12:04.400 "data_offset": 0, 00:12:04.400 "data_size": 63488 00:12:04.400 }, 00:12:04.400 { 00:12:04.400 "name": "BaseBdev2", 00:12:04.400 "uuid": "477c21a6-2fad-53aa-8334-14a66e35f763", 00:12:04.400 "is_configured": true, 00:12:04.400 "data_offset": 2048, 00:12:04.400 "data_size": 63488 00:12:04.400 } 00:12:04.400 ] 00:12:04.400 }' 00:12:04.400 15:16:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:04.400 15:16:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:04.400 15:16:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:04.400 15:16:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:04.400 15:16:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:04.400 15:16:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:04.400 15:16:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:04.400 [2024-11-27 15:16:32.370451] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:04.400 15:16:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:04.400 15:16:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:12:04.400 [2024-11-27 15:16:32.421778] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:12:04.400 [2024-11-27 15:16:32.424230] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:04.660 [2024-11-27 15:16:32.537973] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:04.660 [2024-11-27 15:16:32.538823] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:04.660 [2024-11-27 15:16:32.753485] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:04.660 [2024-11-27 15:16:32.754075] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:05.229 164.67 IOPS, 494.00 MiB/s [2024-11-27T15:16:33.336Z] [2024-11-27 15:16:33.243211] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:05.229 [2024-11-27 15:16:33.243692] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:05.489 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:05.489 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:05.489 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:05.489 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:05.489 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:05.489 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:05.489 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:05.489 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:05.489 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:05.489 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:05.489 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:05.489 "name": "raid_bdev1", 00:12:05.489 "uuid": "ff205561-c290-495f-8446-929b386e84dd", 00:12:05.489 "strip_size_kb": 0, 00:12:05.489 "state": "online", 00:12:05.489 "raid_level": "raid1", 00:12:05.489 "superblock": true, 00:12:05.489 "num_base_bdevs": 2, 00:12:05.489 "num_base_bdevs_discovered": 2, 00:12:05.489 "num_base_bdevs_operational": 2, 00:12:05.489 "process": { 00:12:05.489 "type": "rebuild", 00:12:05.489 "target": "spare", 00:12:05.489 "progress": { 00:12:05.489 "blocks": 12288, 00:12:05.489 "percent": 19 00:12:05.489 } 00:12:05.489 }, 00:12:05.489 "base_bdevs_list": [ 00:12:05.489 { 00:12:05.489 "name": "spare", 00:12:05.489 "uuid": "5ff3b68f-1f8b-5b9b-b9e3-6a342f601efb", 00:12:05.489 "is_configured": true, 00:12:05.489 "data_offset": 2048, 00:12:05.489 "data_size": 63488 00:12:05.489 }, 00:12:05.489 { 00:12:05.489 "name": "BaseBdev2", 00:12:05.489 "uuid": "477c21a6-2fad-53aa-8334-14a66e35f763", 00:12:05.489 "is_configured": true, 00:12:05.489 "data_offset": 2048, 00:12:05.489 "data_size": 63488 00:12:05.489 } 00:12:05.489 ] 00:12:05.489 }' 00:12:05.489 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:05.489 [2024-11-27 15:16:33.475666] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:12:05.489 [2024-11-27 15:16:33.476156] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:12:05.489 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:05.489 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:05.489 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:05.489 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:12:05.489 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:12:05.489 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:12:05.489 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:12:05.489 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:12:05.489 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:12:05.489 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@706 -- # local timeout=334 00:12:05.489 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:05.489 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:05.489 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:05.489 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:05.489 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:05.489 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:05.489 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:05.489 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:05.489 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:05.489 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:05.489 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:05.749 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:05.749 "name": "raid_bdev1", 00:12:05.749 "uuid": "ff205561-c290-495f-8446-929b386e84dd", 00:12:05.749 "strip_size_kb": 0, 00:12:05.749 "state": "online", 00:12:05.749 "raid_level": "raid1", 00:12:05.749 "superblock": true, 00:12:05.749 "num_base_bdevs": 2, 00:12:05.749 "num_base_bdevs_discovered": 2, 00:12:05.749 "num_base_bdevs_operational": 2, 00:12:05.749 "process": { 00:12:05.749 "type": "rebuild", 00:12:05.749 "target": "spare", 00:12:05.749 "progress": { 00:12:05.749 "blocks": 14336, 00:12:05.749 "percent": 22 00:12:05.749 } 00:12:05.749 }, 00:12:05.749 "base_bdevs_list": [ 00:12:05.749 { 00:12:05.749 "name": "spare", 00:12:05.749 "uuid": "5ff3b68f-1f8b-5b9b-b9e3-6a342f601efb", 00:12:05.749 "is_configured": true, 00:12:05.749 "data_offset": 2048, 00:12:05.749 "data_size": 63488 00:12:05.749 }, 00:12:05.749 { 00:12:05.749 "name": "BaseBdev2", 00:12:05.749 "uuid": "477c21a6-2fad-53aa-8334-14a66e35f763", 00:12:05.749 "is_configured": true, 00:12:05.749 "data_offset": 2048, 00:12:05.749 "data_size": 63488 00:12:05.749 } 00:12:05.749 ] 00:12:05.749 }' 00:12:05.749 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:05.749 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:05.749 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:05.749 [2024-11-27 15:16:33.702021] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:12:05.749 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:05.749 15:16:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:06.008 [2024-11-27 15:16:34.033038] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:12:06.008 [2024-11-27 15:16:34.033612] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:12:06.267 146.75 IOPS, 440.25 MiB/s [2024-11-27T15:16:34.374Z] [2024-11-27 15:16:34.162950] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:12:06.526 [2024-11-27 15:16:34.383787] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:12:06.526 [2024-11-27 15:16:34.384743] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:12:06.526 [2024-11-27 15:16:34.516669] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 28672 offset_begin: 24576 offset_end: 30720 00:12:06.786 15:16:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:06.786 15:16:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:06.786 15:16:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:06.786 15:16:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:06.786 15:16:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:06.786 15:16:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:06.786 15:16:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:06.786 15:16:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:06.786 15:16:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:06.786 15:16:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:06.786 15:16:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:06.786 15:16:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:06.786 "name": "raid_bdev1", 00:12:06.786 "uuid": "ff205561-c290-495f-8446-929b386e84dd", 00:12:06.786 "strip_size_kb": 0, 00:12:06.786 "state": "online", 00:12:06.786 "raid_level": "raid1", 00:12:06.786 "superblock": true, 00:12:06.786 "num_base_bdevs": 2, 00:12:06.786 "num_base_bdevs_discovered": 2, 00:12:06.786 "num_base_bdevs_operational": 2, 00:12:06.786 "process": { 00:12:06.786 "type": "rebuild", 00:12:06.786 "target": "spare", 00:12:06.786 "progress": { 00:12:06.786 "blocks": 28672, 00:12:06.786 "percent": 45 00:12:06.786 } 00:12:06.786 }, 00:12:06.786 "base_bdevs_list": [ 00:12:06.786 { 00:12:06.786 "name": "spare", 00:12:06.786 "uuid": "5ff3b68f-1f8b-5b9b-b9e3-6a342f601efb", 00:12:06.786 "is_configured": true, 00:12:06.786 "data_offset": 2048, 00:12:06.786 "data_size": 63488 00:12:06.786 }, 00:12:06.786 { 00:12:06.786 "name": "BaseBdev2", 00:12:06.786 "uuid": "477c21a6-2fad-53aa-8334-14a66e35f763", 00:12:06.786 "is_configured": true, 00:12:06.786 "data_offset": 2048, 00:12:06.786 "data_size": 63488 00:12:06.786 } 00:12:06.786 ] 00:12:06.786 }' 00:12:06.786 15:16:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:06.786 15:16:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:06.786 15:16:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:06.786 15:16:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:06.786 15:16:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:07.609 126.00 IOPS, 378.00 MiB/s [2024-11-27T15:16:35.716Z] [2024-11-27 15:16:35.605005] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 47104 offset_begin: 43008 offset_end: 49152 00:12:07.868 15:16:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:07.868 15:16:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:07.868 15:16:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:07.869 15:16:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:07.869 15:16:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:07.869 15:16:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:07.869 15:16:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:07.869 15:16:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:07.869 15:16:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:07.869 15:16:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:07.869 15:16:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:07.869 15:16:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:07.869 "name": "raid_bdev1", 00:12:07.869 "uuid": "ff205561-c290-495f-8446-929b386e84dd", 00:12:07.869 "strip_size_kb": 0, 00:12:07.869 "state": "online", 00:12:07.869 "raid_level": "raid1", 00:12:07.869 "superblock": true, 00:12:07.869 "num_base_bdevs": 2, 00:12:07.869 "num_base_bdevs_discovered": 2, 00:12:07.869 "num_base_bdevs_operational": 2, 00:12:07.869 "process": { 00:12:07.869 "type": "rebuild", 00:12:07.869 "target": "spare", 00:12:07.869 "progress": { 00:12:07.869 "blocks": 51200, 00:12:07.869 "percent": 80 00:12:07.869 } 00:12:07.869 }, 00:12:07.869 "base_bdevs_list": [ 00:12:07.869 { 00:12:07.869 "name": "spare", 00:12:07.869 "uuid": "5ff3b68f-1f8b-5b9b-b9e3-6a342f601efb", 00:12:07.869 "is_configured": true, 00:12:07.869 "data_offset": 2048, 00:12:07.869 "data_size": 63488 00:12:07.869 }, 00:12:07.869 { 00:12:07.869 "name": "BaseBdev2", 00:12:07.869 "uuid": "477c21a6-2fad-53aa-8334-14a66e35f763", 00:12:07.869 "is_configured": true, 00:12:07.869 "data_offset": 2048, 00:12:07.869 "data_size": 63488 00:12:07.869 } 00:12:07.869 ] 00:12:07.869 }' 00:12:07.869 15:16:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:07.869 15:16:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:07.869 15:16:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:08.129 15:16:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:08.129 15:16:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:08.697 112.00 IOPS, 336.00 MiB/s [2024-11-27T15:16:36.804Z] [2024-11-27 15:16:36.526355] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:12:08.697 [2024-11-27 15:16:36.631463] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:12:08.697 [2024-11-27 15:16:36.635867] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:08.956 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:08.956 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:08.956 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:08.956 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:08.956 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:08.956 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:08.956 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:08.956 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:08.956 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.956 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:08.956 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.956 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:08.956 "name": "raid_bdev1", 00:12:08.956 "uuid": "ff205561-c290-495f-8446-929b386e84dd", 00:12:08.956 "strip_size_kb": 0, 00:12:08.956 "state": "online", 00:12:08.956 "raid_level": "raid1", 00:12:08.956 "superblock": true, 00:12:08.956 "num_base_bdevs": 2, 00:12:08.956 "num_base_bdevs_discovered": 2, 00:12:08.956 "num_base_bdevs_operational": 2, 00:12:08.956 "base_bdevs_list": [ 00:12:08.956 { 00:12:08.956 "name": "spare", 00:12:08.956 "uuid": "5ff3b68f-1f8b-5b9b-b9e3-6a342f601efb", 00:12:08.956 "is_configured": true, 00:12:08.956 "data_offset": 2048, 00:12:08.956 "data_size": 63488 00:12:08.956 }, 00:12:08.956 { 00:12:08.956 "name": "BaseBdev2", 00:12:08.956 "uuid": "477c21a6-2fad-53aa-8334-14a66e35f763", 00:12:08.956 "is_configured": true, 00:12:08.956 "data_offset": 2048, 00:12:08.956 "data_size": 63488 00:12:08.956 } 00:12:08.956 ] 00:12:08.956 }' 00:12:08.956 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:09.216 99.57 IOPS, 298.71 MiB/s [2024-11-27T15:16:37.323Z] 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@709 -- # break 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:09.216 "name": "raid_bdev1", 00:12:09.216 "uuid": "ff205561-c290-495f-8446-929b386e84dd", 00:12:09.216 "strip_size_kb": 0, 00:12:09.216 "state": "online", 00:12:09.216 "raid_level": "raid1", 00:12:09.216 "superblock": true, 00:12:09.216 "num_base_bdevs": 2, 00:12:09.216 "num_base_bdevs_discovered": 2, 00:12:09.216 "num_base_bdevs_operational": 2, 00:12:09.216 "base_bdevs_list": [ 00:12:09.216 { 00:12:09.216 "name": "spare", 00:12:09.216 "uuid": "5ff3b68f-1f8b-5b9b-b9e3-6a342f601efb", 00:12:09.216 "is_configured": true, 00:12:09.216 "data_offset": 2048, 00:12:09.216 "data_size": 63488 00:12:09.216 }, 00:12:09.216 { 00:12:09.216 "name": "BaseBdev2", 00:12:09.216 "uuid": "477c21a6-2fad-53aa-8334-14a66e35f763", 00:12:09.216 "is_configured": true, 00:12:09.216 "data_offset": 2048, 00:12:09.216 "data_size": 63488 00:12:09.216 } 00:12:09.216 ] 00:12:09.216 }' 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:09.216 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.475 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:09.475 "name": "raid_bdev1", 00:12:09.475 "uuid": "ff205561-c290-495f-8446-929b386e84dd", 00:12:09.475 "strip_size_kb": 0, 00:12:09.475 "state": "online", 00:12:09.475 "raid_level": "raid1", 00:12:09.475 "superblock": true, 00:12:09.475 "num_base_bdevs": 2, 00:12:09.475 "num_base_bdevs_discovered": 2, 00:12:09.475 "num_base_bdevs_operational": 2, 00:12:09.475 "base_bdevs_list": [ 00:12:09.475 { 00:12:09.475 "name": "spare", 00:12:09.475 "uuid": "5ff3b68f-1f8b-5b9b-b9e3-6a342f601efb", 00:12:09.475 "is_configured": true, 00:12:09.475 "data_offset": 2048, 00:12:09.475 "data_size": 63488 00:12:09.475 }, 00:12:09.475 { 00:12:09.475 "name": "BaseBdev2", 00:12:09.475 "uuid": "477c21a6-2fad-53aa-8334-14a66e35f763", 00:12:09.475 "is_configured": true, 00:12:09.475 "data_offset": 2048, 00:12:09.475 "data_size": 63488 00:12:09.475 } 00:12:09.475 ] 00:12:09.475 }' 00:12:09.475 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:09.475 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:09.734 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:09.734 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.734 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:09.734 [2024-11-27 15:16:37.730996] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:09.734 [2024-11-27 15:16:37.731137] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:09.734 00:12:09.734 Latency(us) 00:12:09.734 [2024-11-27T15:16:37.841Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:09.734 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:12:09.734 raid_bdev1 : 7.73 93.70 281.11 0.00 0.00 15073.17 287.97 114473.36 00:12:09.734 [2024-11-27T15:16:37.841Z] =================================================================================================================== 00:12:09.734 [2024-11-27T15:16:37.841Z] Total : 93.70 281.11 0.00 0.00 15073.17 287.97 114473.36 00:12:09.734 [2024-11-27 15:16:37.823007] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:09.734 [2024-11-27 15:16:37.823126] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:09.734 [2024-11-27 15:16:37.823227] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:09.734 [2024-11-27 15:16:37.823300] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:12:09.734 { 00:12:09.734 "results": [ 00:12:09.734 { 00:12:09.734 "job": "raid_bdev1", 00:12:09.734 "core_mask": "0x1", 00:12:09.734 "workload": "randrw", 00:12:09.734 "percentage": 50, 00:12:09.734 "status": "finished", 00:12:09.734 "queue_depth": 2, 00:12:09.734 "io_size": 3145728, 00:12:09.734 "runtime": 7.726475, 00:12:09.734 "iops": 93.7037911855018, 00:12:09.734 "mibps": 281.1113735565054, 00:12:09.734 "io_failed": 0, 00:12:09.734 "io_timeout": 0, 00:12:09.734 "avg_latency_us": 15073.174262346496, 00:12:09.734 "min_latency_us": 287.97205240174674, 00:12:09.734 "max_latency_us": 114473.36244541485 00:12:09.734 } 00:12:09.734 ], 00:12:09.734 "core_count": 1 00:12:09.734 } 00:12:09.734 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.734 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:09.734 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.734 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # jq length 00:12:09.734 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:09.993 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.993 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:12:09.993 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:12:09.993 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:12:09.993 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:12:09.993 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:09.993 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:12:09.993 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:09.993 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:12:09.993 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:09.993 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:12:09.993 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:09.993 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:09.993 15:16:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:12:09.993 /dev/nbd0 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:10.253 1+0 records in 00:12:10.253 1+0 records out 00:12:10.253 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000583642 s, 7.0 MB/s 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev2 ']' 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev2 /dev/nbd1 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev2') 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev2 /dev/nbd1 00:12:10.253 /dev/nbd1 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:10.253 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:12:10.512 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:12:10.512 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:10.512 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:10.512 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:10.512 1+0 records in 00:12:10.512 1+0 records out 00:12:10.512 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000291106 s, 14.1 MB/s 00:12:10.512 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:10.512 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:12:10.512 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:10.512 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:10.512 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:12:10.512 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:10.512 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:10.512 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:12:10.512 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:12:10.512 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:10.512 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:12:10.512 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:10.512 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:12:10.512 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:10.512 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:10.772 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:10.772 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:10.772 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:10.772 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:10.772 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:10.772 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:10.772 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:12:10.772 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:10.772 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:12:10.772 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:10.772 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:12:10.772 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:10.772 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:12:10.772 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:10.772 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:10.772 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:10.772 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:10.772 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:10.772 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:10.772 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:10.772 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:10.772 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:12:10.772 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:10.772 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:12:10.772 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:12:10.772 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:10.772 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:10.772 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:10.772 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:10.772 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:10.772 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:10.772 [2024-11-27 15:16:38.872123] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:10.772 [2024-11-27 15:16:38.872200] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:10.772 [2024-11-27 15:16:38.872227] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:12:10.772 [2024-11-27 15:16:38.872241] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:10.772 [2024-11-27 15:16:38.874798] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:10.772 [2024-11-27 15:16:38.874850] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:10.772 [2024-11-27 15:16:38.874978] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:12:10.772 [2024-11-27 15:16:38.875035] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:10.772 [2024-11-27 15:16:38.875186] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:11.031 spare 00:12:11.031 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.031 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:12:11.031 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.031 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:11.031 [2024-11-27 15:16:38.975110] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006600 00:12:11.031 [2024-11-27 15:16:38.975217] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:11.031 [2024-11-27 15:16:38.975551] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002af30 00:12:11.031 [2024-11-27 15:16:38.975734] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006600 00:12:11.031 [2024-11-27 15:16:38.975753] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006600 00:12:11.031 [2024-11-27 15:16:38.975966] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:11.031 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.031 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:11.031 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:11.031 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:11.031 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:11.031 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:11.031 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:11.031 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:11.031 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:11.031 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:11.032 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:11.032 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:11.032 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.032 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:11.032 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:11.032 15:16:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.032 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:11.032 "name": "raid_bdev1", 00:12:11.032 "uuid": "ff205561-c290-495f-8446-929b386e84dd", 00:12:11.032 "strip_size_kb": 0, 00:12:11.032 "state": "online", 00:12:11.032 "raid_level": "raid1", 00:12:11.032 "superblock": true, 00:12:11.032 "num_base_bdevs": 2, 00:12:11.032 "num_base_bdevs_discovered": 2, 00:12:11.032 "num_base_bdevs_operational": 2, 00:12:11.032 "base_bdevs_list": [ 00:12:11.032 { 00:12:11.032 "name": "spare", 00:12:11.032 "uuid": "5ff3b68f-1f8b-5b9b-b9e3-6a342f601efb", 00:12:11.032 "is_configured": true, 00:12:11.032 "data_offset": 2048, 00:12:11.032 "data_size": 63488 00:12:11.032 }, 00:12:11.032 { 00:12:11.032 "name": "BaseBdev2", 00:12:11.032 "uuid": "477c21a6-2fad-53aa-8334-14a66e35f763", 00:12:11.032 "is_configured": true, 00:12:11.032 "data_offset": 2048, 00:12:11.032 "data_size": 63488 00:12:11.032 } 00:12:11.032 ] 00:12:11.032 }' 00:12:11.032 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:11.032 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:11.598 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:11.598 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:11.598 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:11.598 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:11.598 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:11.598 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:11.598 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:11.598 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.598 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:11.598 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.598 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:11.598 "name": "raid_bdev1", 00:12:11.598 "uuid": "ff205561-c290-495f-8446-929b386e84dd", 00:12:11.598 "strip_size_kb": 0, 00:12:11.598 "state": "online", 00:12:11.598 "raid_level": "raid1", 00:12:11.598 "superblock": true, 00:12:11.598 "num_base_bdevs": 2, 00:12:11.598 "num_base_bdevs_discovered": 2, 00:12:11.598 "num_base_bdevs_operational": 2, 00:12:11.598 "base_bdevs_list": [ 00:12:11.598 { 00:12:11.598 "name": "spare", 00:12:11.598 "uuid": "5ff3b68f-1f8b-5b9b-b9e3-6a342f601efb", 00:12:11.598 "is_configured": true, 00:12:11.598 "data_offset": 2048, 00:12:11.598 "data_size": 63488 00:12:11.598 }, 00:12:11.598 { 00:12:11.599 "name": "BaseBdev2", 00:12:11.599 "uuid": "477c21a6-2fad-53aa-8334-14a66e35f763", 00:12:11.599 "is_configured": true, 00:12:11.599 "data_offset": 2048, 00:12:11.599 "data_size": 63488 00:12:11.599 } 00:12:11.599 ] 00:12:11.599 }' 00:12:11.599 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:11.599 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:11.599 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:11.599 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:11.599 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:11.599 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.599 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:11.599 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:12:11.599 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.599 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:12:11.599 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:11.599 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.599 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:11.599 [2024-11-27 15:16:39.635037] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:11.599 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.599 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:11.599 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:11.599 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:11.599 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:11.599 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:11.599 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:11.599 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:11.599 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:11.599 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:11.599 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:11.599 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:11.599 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.599 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:11.599 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:11.599 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.599 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:11.599 "name": "raid_bdev1", 00:12:11.599 "uuid": "ff205561-c290-495f-8446-929b386e84dd", 00:12:11.599 "strip_size_kb": 0, 00:12:11.599 "state": "online", 00:12:11.599 "raid_level": "raid1", 00:12:11.599 "superblock": true, 00:12:11.599 "num_base_bdevs": 2, 00:12:11.599 "num_base_bdevs_discovered": 1, 00:12:11.599 "num_base_bdevs_operational": 1, 00:12:11.599 "base_bdevs_list": [ 00:12:11.599 { 00:12:11.599 "name": null, 00:12:11.599 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:11.599 "is_configured": false, 00:12:11.599 "data_offset": 0, 00:12:11.599 "data_size": 63488 00:12:11.599 }, 00:12:11.599 { 00:12:11.599 "name": "BaseBdev2", 00:12:11.599 "uuid": "477c21a6-2fad-53aa-8334-14a66e35f763", 00:12:11.599 "is_configured": true, 00:12:11.599 "data_offset": 2048, 00:12:11.599 "data_size": 63488 00:12:11.599 } 00:12:11.599 ] 00:12:11.599 }' 00:12:11.599 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:11.599 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:12.168 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:12.168 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:12.168 15:16:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:12.168 [2024-11-27 15:16:40.006509] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:12.168 [2024-11-27 15:16:40.006800] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:12:12.169 [2024-11-27 15:16:40.006818] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:12:12.169 [2024-11-27 15:16:40.006887] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:12.169 [2024-11-27 15:16:40.016383] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b000 00:12:12.169 15:16:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:12.169 15:16:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@757 -- # sleep 1 00:12:12.169 [2024-11-27 15:16:40.018605] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:13.111 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:13.111 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:13.111 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:13.111 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:13.111 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:13.111 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:13.111 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:13.111 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:13.111 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:13.111 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:13.111 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:13.111 "name": "raid_bdev1", 00:12:13.111 "uuid": "ff205561-c290-495f-8446-929b386e84dd", 00:12:13.111 "strip_size_kb": 0, 00:12:13.111 "state": "online", 00:12:13.111 "raid_level": "raid1", 00:12:13.111 "superblock": true, 00:12:13.111 "num_base_bdevs": 2, 00:12:13.111 "num_base_bdevs_discovered": 2, 00:12:13.111 "num_base_bdevs_operational": 2, 00:12:13.111 "process": { 00:12:13.111 "type": "rebuild", 00:12:13.111 "target": "spare", 00:12:13.111 "progress": { 00:12:13.111 "blocks": 20480, 00:12:13.111 "percent": 32 00:12:13.111 } 00:12:13.111 }, 00:12:13.111 "base_bdevs_list": [ 00:12:13.111 { 00:12:13.111 "name": "spare", 00:12:13.111 "uuid": "5ff3b68f-1f8b-5b9b-b9e3-6a342f601efb", 00:12:13.111 "is_configured": true, 00:12:13.111 "data_offset": 2048, 00:12:13.111 "data_size": 63488 00:12:13.111 }, 00:12:13.111 { 00:12:13.111 "name": "BaseBdev2", 00:12:13.111 "uuid": "477c21a6-2fad-53aa-8334-14a66e35f763", 00:12:13.111 "is_configured": true, 00:12:13.111 "data_offset": 2048, 00:12:13.111 "data_size": 63488 00:12:13.111 } 00:12:13.111 ] 00:12:13.111 }' 00:12:13.111 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:13.111 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:13.111 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:13.111 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:13.111 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:12:13.111 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:13.111 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:13.111 [2024-11-27 15:16:41.179313] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:13.371 [2024-11-27 15:16:41.227154] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:13.371 [2024-11-27 15:16:41.227227] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:13.371 [2024-11-27 15:16:41.227249] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:13.371 [2024-11-27 15:16:41.227258] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:13.371 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:13.371 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:13.371 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:13.371 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:13.371 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:13.371 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:13.371 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:13.371 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:13.371 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:13.371 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:13.371 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:13.371 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:13.371 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:13.371 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:13.371 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:13.371 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:13.371 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:13.371 "name": "raid_bdev1", 00:12:13.371 "uuid": "ff205561-c290-495f-8446-929b386e84dd", 00:12:13.371 "strip_size_kb": 0, 00:12:13.371 "state": "online", 00:12:13.371 "raid_level": "raid1", 00:12:13.371 "superblock": true, 00:12:13.371 "num_base_bdevs": 2, 00:12:13.371 "num_base_bdevs_discovered": 1, 00:12:13.371 "num_base_bdevs_operational": 1, 00:12:13.371 "base_bdevs_list": [ 00:12:13.371 { 00:12:13.371 "name": null, 00:12:13.371 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:13.371 "is_configured": false, 00:12:13.371 "data_offset": 0, 00:12:13.371 "data_size": 63488 00:12:13.371 }, 00:12:13.371 { 00:12:13.371 "name": "BaseBdev2", 00:12:13.371 "uuid": "477c21a6-2fad-53aa-8334-14a66e35f763", 00:12:13.371 "is_configured": true, 00:12:13.371 "data_offset": 2048, 00:12:13.371 "data_size": 63488 00:12:13.371 } 00:12:13.371 ] 00:12:13.371 }' 00:12:13.371 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:13.371 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:13.631 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:13.631 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:13.631 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:13.631 [2024-11-27 15:16:41.703590] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:13.631 [2024-11-27 15:16:41.703786] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:13.631 [2024-11-27 15:16:41.703847] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:12:13.631 [2024-11-27 15:16:41.703883] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:13.631 [2024-11-27 15:16:41.704528] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:13.631 [2024-11-27 15:16:41.704611] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:13.631 [2024-11-27 15:16:41.704764] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:12:13.631 [2024-11-27 15:16:41.704811] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:12:13.631 [2024-11-27 15:16:41.704867] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:12:13.631 [2024-11-27 15:16:41.704961] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:13.631 [2024-11-27 15:16:41.714435] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b0d0 00:12:13.631 spare 00:12:13.631 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:13.631 15:16:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@764 -- # sleep 1 00:12:13.631 [2024-11-27 15:16:41.716806] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:15.010 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:15.010 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:15.010 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:15.010 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:15.010 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:15.010 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:15.010 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:15.010 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:15.010 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:15.010 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:15.010 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:15.010 "name": "raid_bdev1", 00:12:15.010 "uuid": "ff205561-c290-495f-8446-929b386e84dd", 00:12:15.010 "strip_size_kb": 0, 00:12:15.010 "state": "online", 00:12:15.010 "raid_level": "raid1", 00:12:15.010 "superblock": true, 00:12:15.010 "num_base_bdevs": 2, 00:12:15.010 "num_base_bdevs_discovered": 2, 00:12:15.010 "num_base_bdevs_operational": 2, 00:12:15.010 "process": { 00:12:15.010 "type": "rebuild", 00:12:15.010 "target": "spare", 00:12:15.010 "progress": { 00:12:15.010 "blocks": 20480, 00:12:15.010 "percent": 32 00:12:15.010 } 00:12:15.010 }, 00:12:15.010 "base_bdevs_list": [ 00:12:15.010 { 00:12:15.010 "name": "spare", 00:12:15.010 "uuid": "5ff3b68f-1f8b-5b9b-b9e3-6a342f601efb", 00:12:15.010 "is_configured": true, 00:12:15.010 "data_offset": 2048, 00:12:15.010 "data_size": 63488 00:12:15.010 }, 00:12:15.010 { 00:12:15.010 "name": "BaseBdev2", 00:12:15.010 "uuid": "477c21a6-2fad-53aa-8334-14a66e35f763", 00:12:15.010 "is_configured": true, 00:12:15.010 "data_offset": 2048, 00:12:15.010 "data_size": 63488 00:12:15.010 } 00:12:15.010 ] 00:12:15.010 }' 00:12:15.010 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:15.010 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:15.010 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:15.010 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:15.010 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:12:15.010 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:15.010 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:15.010 [2024-11-27 15:16:42.877284] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:15.010 [2024-11-27 15:16:42.925990] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:15.010 [2024-11-27 15:16:42.926087] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:15.010 [2024-11-27 15:16:42.926106] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:15.010 [2024-11-27 15:16:42.926122] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:15.010 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:15.010 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:15.010 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:15.010 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:15.010 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:15.010 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:15.011 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:15.011 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:15.011 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:15.011 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:15.011 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:15.011 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:15.011 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:15.011 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:15.011 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:15.011 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:15.011 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:15.011 "name": "raid_bdev1", 00:12:15.011 "uuid": "ff205561-c290-495f-8446-929b386e84dd", 00:12:15.011 "strip_size_kb": 0, 00:12:15.011 "state": "online", 00:12:15.011 "raid_level": "raid1", 00:12:15.011 "superblock": true, 00:12:15.011 "num_base_bdevs": 2, 00:12:15.011 "num_base_bdevs_discovered": 1, 00:12:15.011 "num_base_bdevs_operational": 1, 00:12:15.011 "base_bdevs_list": [ 00:12:15.011 { 00:12:15.011 "name": null, 00:12:15.011 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:15.011 "is_configured": false, 00:12:15.011 "data_offset": 0, 00:12:15.011 "data_size": 63488 00:12:15.011 }, 00:12:15.011 { 00:12:15.011 "name": "BaseBdev2", 00:12:15.011 "uuid": "477c21a6-2fad-53aa-8334-14a66e35f763", 00:12:15.011 "is_configured": true, 00:12:15.011 "data_offset": 2048, 00:12:15.011 "data_size": 63488 00:12:15.011 } 00:12:15.011 ] 00:12:15.011 }' 00:12:15.011 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:15.011 15:16:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:15.605 15:16:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:15.606 15:16:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:15.606 15:16:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:15.606 15:16:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:15.606 15:16:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:15.606 15:16:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:15.606 15:16:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:15.606 15:16:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:15.606 15:16:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:15.606 15:16:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:15.606 15:16:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:15.606 "name": "raid_bdev1", 00:12:15.606 "uuid": "ff205561-c290-495f-8446-929b386e84dd", 00:12:15.606 "strip_size_kb": 0, 00:12:15.606 "state": "online", 00:12:15.606 "raid_level": "raid1", 00:12:15.606 "superblock": true, 00:12:15.606 "num_base_bdevs": 2, 00:12:15.606 "num_base_bdevs_discovered": 1, 00:12:15.606 "num_base_bdevs_operational": 1, 00:12:15.606 "base_bdevs_list": [ 00:12:15.606 { 00:12:15.606 "name": null, 00:12:15.606 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:15.606 "is_configured": false, 00:12:15.606 "data_offset": 0, 00:12:15.606 "data_size": 63488 00:12:15.606 }, 00:12:15.606 { 00:12:15.606 "name": "BaseBdev2", 00:12:15.606 "uuid": "477c21a6-2fad-53aa-8334-14a66e35f763", 00:12:15.606 "is_configured": true, 00:12:15.606 "data_offset": 2048, 00:12:15.606 "data_size": 63488 00:12:15.606 } 00:12:15.606 ] 00:12:15.606 }' 00:12:15.606 15:16:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:15.606 15:16:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:15.606 15:16:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:15.606 15:16:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:15.606 15:16:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:12:15.606 15:16:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:15.606 15:16:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:15.606 15:16:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:15.606 15:16:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:15.606 15:16:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:15.606 15:16:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:15.606 [2024-11-27 15:16:43.598117] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:15.606 [2024-11-27 15:16:43.598195] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:15.606 [2024-11-27 15:16:43.598221] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:12:15.606 [2024-11-27 15:16:43.598236] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:15.606 [2024-11-27 15:16:43.598746] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:15.606 [2024-11-27 15:16:43.598787] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:15.606 [2024-11-27 15:16:43.598879] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:12:15.606 [2024-11-27 15:16:43.598900] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:12:15.606 [2024-11-27 15:16:43.598933] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:12:15.606 [2024-11-27 15:16:43.598952] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:12:15.606 BaseBdev1 00:12:15.606 15:16:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:15.606 15:16:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@775 -- # sleep 1 00:12:16.546 15:16:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:16.546 15:16:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:16.546 15:16:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:16.546 15:16:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:16.546 15:16:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:16.546 15:16:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:16.546 15:16:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:16.546 15:16:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:16.546 15:16:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:16.546 15:16:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:16.546 15:16:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:16.546 15:16:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:16.546 15:16:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:16.546 15:16:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:16.546 15:16:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:16.806 15:16:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:16.806 "name": "raid_bdev1", 00:12:16.806 "uuid": "ff205561-c290-495f-8446-929b386e84dd", 00:12:16.806 "strip_size_kb": 0, 00:12:16.806 "state": "online", 00:12:16.806 "raid_level": "raid1", 00:12:16.806 "superblock": true, 00:12:16.806 "num_base_bdevs": 2, 00:12:16.806 "num_base_bdevs_discovered": 1, 00:12:16.806 "num_base_bdevs_operational": 1, 00:12:16.806 "base_bdevs_list": [ 00:12:16.806 { 00:12:16.806 "name": null, 00:12:16.806 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:16.806 "is_configured": false, 00:12:16.806 "data_offset": 0, 00:12:16.806 "data_size": 63488 00:12:16.806 }, 00:12:16.806 { 00:12:16.806 "name": "BaseBdev2", 00:12:16.806 "uuid": "477c21a6-2fad-53aa-8334-14a66e35f763", 00:12:16.806 "is_configured": true, 00:12:16.806 "data_offset": 2048, 00:12:16.806 "data_size": 63488 00:12:16.806 } 00:12:16.806 ] 00:12:16.806 }' 00:12:16.806 15:16:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:16.806 15:16:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:17.065 15:16:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:17.065 15:16:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:17.065 15:16:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:17.065 15:16:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:17.065 15:16:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:17.065 15:16:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:17.065 15:16:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:17.065 15:16:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:17.065 15:16:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:17.065 15:16:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:17.065 15:16:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:17.065 "name": "raid_bdev1", 00:12:17.065 "uuid": "ff205561-c290-495f-8446-929b386e84dd", 00:12:17.065 "strip_size_kb": 0, 00:12:17.065 "state": "online", 00:12:17.065 "raid_level": "raid1", 00:12:17.065 "superblock": true, 00:12:17.065 "num_base_bdevs": 2, 00:12:17.065 "num_base_bdevs_discovered": 1, 00:12:17.065 "num_base_bdevs_operational": 1, 00:12:17.065 "base_bdevs_list": [ 00:12:17.065 { 00:12:17.065 "name": null, 00:12:17.065 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:17.065 "is_configured": false, 00:12:17.065 "data_offset": 0, 00:12:17.065 "data_size": 63488 00:12:17.065 }, 00:12:17.065 { 00:12:17.065 "name": "BaseBdev2", 00:12:17.065 "uuid": "477c21a6-2fad-53aa-8334-14a66e35f763", 00:12:17.065 "is_configured": true, 00:12:17.065 "data_offset": 2048, 00:12:17.065 "data_size": 63488 00:12:17.065 } 00:12:17.065 ] 00:12:17.065 }' 00:12:17.065 15:16:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:17.325 15:16:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:17.325 15:16:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:17.325 15:16:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:17.325 15:16:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:17.325 15:16:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@652 -- # local es=0 00:12:17.325 15:16:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:17.325 15:16:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:12:17.325 15:16:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:12:17.325 15:16:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:12:17.325 15:16:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:12:17.325 15:16:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:17.325 15:16:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:17.325 15:16:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:17.325 [2024-11-27 15:16:45.235708] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:17.325 [2024-11-27 15:16:45.236041] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:12:17.325 [2024-11-27 15:16:45.236063] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:12:17.325 request: 00:12:17.325 { 00:12:17.325 "base_bdev": "BaseBdev1", 00:12:17.325 "raid_bdev": "raid_bdev1", 00:12:17.325 "method": "bdev_raid_add_base_bdev", 00:12:17.325 "req_id": 1 00:12:17.325 } 00:12:17.325 Got JSON-RPC error response 00:12:17.325 response: 00:12:17.325 { 00:12:17.325 "code": -22, 00:12:17.325 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:12:17.325 } 00:12:17.325 15:16:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:12:17.325 15:16:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@655 -- # es=1 00:12:17.325 15:16:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:12:17.325 15:16:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:12:17.325 15:16:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:12:17.325 15:16:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@779 -- # sleep 1 00:12:18.263 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:18.263 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:18.263 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:18.263 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:18.263 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:18.263 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:18.263 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:18.263 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:18.263 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:18.263 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:18.263 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:18.263 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:18.263 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:18.263 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:18.263 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:18.263 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:18.263 "name": "raid_bdev1", 00:12:18.263 "uuid": "ff205561-c290-495f-8446-929b386e84dd", 00:12:18.263 "strip_size_kb": 0, 00:12:18.263 "state": "online", 00:12:18.263 "raid_level": "raid1", 00:12:18.263 "superblock": true, 00:12:18.263 "num_base_bdevs": 2, 00:12:18.263 "num_base_bdevs_discovered": 1, 00:12:18.263 "num_base_bdevs_operational": 1, 00:12:18.263 "base_bdevs_list": [ 00:12:18.263 { 00:12:18.263 "name": null, 00:12:18.263 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:18.263 "is_configured": false, 00:12:18.263 "data_offset": 0, 00:12:18.263 "data_size": 63488 00:12:18.263 }, 00:12:18.263 { 00:12:18.263 "name": "BaseBdev2", 00:12:18.263 "uuid": "477c21a6-2fad-53aa-8334-14a66e35f763", 00:12:18.263 "is_configured": true, 00:12:18.263 "data_offset": 2048, 00:12:18.263 "data_size": 63488 00:12:18.263 } 00:12:18.263 ] 00:12:18.263 }' 00:12:18.263 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:18.263 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:18.836 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:18.836 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:18.836 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:18.836 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:18.836 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:18.836 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:18.836 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:18.836 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:18.836 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:18.836 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:18.836 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:18.836 "name": "raid_bdev1", 00:12:18.836 "uuid": "ff205561-c290-495f-8446-929b386e84dd", 00:12:18.836 "strip_size_kb": 0, 00:12:18.836 "state": "online", 00:12:18.836 "raid_level": "raid1", 00:12:18.836 "superblock": true, 00:12:18.836 "num_base_bdevs": 2, 00:12:18.836 "num_base_bdevs_discovered": 1, 00:12:18.836 "num_base_bdevs_operational": 1, 00:12:18.836 "base_bdevs_list": [ 00:12:18.836 { 00:12:18.836 "name": null, 00:12:18.836 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:18.836 "is_configured": false, 00:12:18.836 "data_offset": 0, 00:12:18.836 "data_size": 63488 00:12:18.836 }, 00:12:18.836 { 00:12:18.836 "name": "BaseBdev2", 00:12:18.836 "uuid": "477c21a6-2fad-53aa-8334-14a66e35f763", 00:12:18.836 "is_configured": true, 00:12:18.836 "data_offset": 2048, 00:12:18.836 "data_size": 63488 00:12:18.836 } 00:12:18.836 ] 00:12:18.836 }' 00:12:18.836 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:18.836 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:18.836 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:18.836 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:18.836 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@784 -- # killprocess 87680 00:12:18.836 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@954 -- # '[' -z 87680 ']' 00:12:18.836 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@958 -- # kill -0 87680 00:12:18.836 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@959 -- # uname 00:12:18.836 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:18.836 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 87680 00:12:18.836 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:18.836 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:18.836 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@972 -- # echo 'killing process with pid 87680' 00:12:18.836 killing process with pid 87680 00:12:18.836 Received shutdown signal, test time was about 16.757773 seconds 00:12:18.836 00:12:18.836 Latency(us) 00:12:18.836 [2024-11-27T15:16:46.943Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:18.836 [2024-11-27T15:16:46.943Z] =================================================================================================================== 00:12:18.836 [2024-11-27T15:16:46.943Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:12:18.836 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@973 -- # kill 87680 00:12:18.836 [2024-11-27 15:16:46.832681] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:18.836 15:16:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@978 -- # wait 87680 00:12:18.836 [2024-11-27 15:16:46.832848] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:18.836 [2024-11-27 15:16:46.832940] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:18.836 [2024-11-27 15:16:46.832953] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state offline 00:12:18.836 [2024-11-27 15:16:46.883414] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@786 -- # return 0 00:12:19.405 00:12:19.405 real 0m18.728s 00:12:19.405 user 0m24.639s 00:12:19.405 sys 0m2.284s 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:19.405 ************************************ 00:12:19.405 END TEST raid_rebuild_test_sb_io 00:12:19.405 ************************************ 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:19.405 15:16:47 bdev_raid -- bdev/bdev_raid.sh@977 -- # for n in 2 4 00:12:19.405 15:16:47 bdev_raid -- bdev/bdev_raid.sh@978 -- # run_test raid_rebuild_test raid_rebuild_test raid1 4 false false true 00:12:19.405 15:16:47 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:12:19.405 15:16:47 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:19.405 15:16:47 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:19.405 ************************************ 00:12:19.405 START TEST raid_rebuild_test 00:12:19.405 ************************************ 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 4 false false true 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=88356 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:19.405 15:16:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 88356 00:12:19.406 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:19.406 15:16:47 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@835 -- # '[' -z 88356 ']' 00:12:19.406 15:16:47 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:19.406 15:16:47 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:19.406 15:16:47 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:19.406 15:16:47 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:19.406 15:16:47 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:19.406 [2024-11-27 15:16:47.382929] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:12:19.406 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:19.406 Zero copy mechanism will not be used. 00:12:19.406 [2024-11-27 15:16:47.383108] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid88356 ] 00:12:19.665 [2024-11-27 15:16:47.550509] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:19.665 [2024-11-27 15:16:47.592665] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:19.665 [2024-11-27 15:16:47.668782] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:19.665 [2024-11-27 15:16:47.669001] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:20.234 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:20.234 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # return 0 00:12:20.234 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:20.234 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:20.234 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:20.234 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:20.234 BaseBdev1_malloc 00:12:20.234 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:20.234 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:20.234 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:20.234 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:20.234 [2024-11-27 15:16:48.317677] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:20.234 [2024-11-27 15:16:48.317754] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:20.234 [2024-11-27 15:16:48.317801] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:20.234 [2024-11-27 15:16:48.317824] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:20.234 [2024-11-27 15:16:48.320343] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:20.234 [2024-11-27 15:16:48.320398] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:20.234 BaseBdev1 00:12:20.234 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:20.234 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:20.234 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:20.234 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:20.234 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:20.494 BaseBdev2_malloc 00:12:20.494 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:20.494 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:12:20.494 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:20.494 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:20.494 [2024-11-27 15:16:48.352211] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:12:20.494 [2024-11-27 15:16:48.352270] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:20.494 [2024-11-27 15:16:48.352295] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:12:20.494 [2024-11-27 15:16:48.352306] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:20.494 [2024-11-27 15:16:48.354653] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:20.494 [2024-11-27 15:16:48.354695] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:20.494 BaseBdev2 00:12:20.494 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:20.494 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:20.494 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:12:20.494 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:20.494 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:20.494 BaseBdev3_malloc 00:12:20.494 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:20.494 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:12:20.494 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:20.494 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:20.494 [2024-11-27 15:16:48.386794] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:12:20.494 [2024-11-27 15:16:48.386944] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:20.494 [2024-11-27 15:16:48.386993] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:12:20.494 [2024-11-27 15:16:48.387026] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:20.494 [2024-11-27 15:16:48.389481] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:20.494 [2024-11-27 15:16:48.389567] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:12:20.494 BaseBdev3 00:12:20.494 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:20.494 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:20.494 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:12:20.494 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:20.494 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:20.494 BaseBdev4_malloc 00:12:20.494 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:20.494 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:12:20.494 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:20.494 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:20.494 [2024-11-27 15:16:48.429220] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:12:20.494 [2024-11-27 15:16:48.429335] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:20.494 [2024-11-27 15:16:48.429384] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:12:20.494 [2024-11-27 15:16:48.429420] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:20.494 [2024-11-27 15:16:48.431723] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:20.494 [2024-11-27 15:16:48.431804] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:12:20.494 BaseBdev4 00:12:20.494 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:20.494 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:12:20.494 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:20.495 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:20.495 spare_malloc 00:12:20.495 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:20.495 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:12:20.495 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:20.495 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:20.495 spare_delay 00:12:20.495 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:20.495 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:20.495 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:20.495 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:20.495 [2024-11-27 15:16:48.475834] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:20.495 [2024-11-27 15:16:48.475887] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:20.495 [2024-11-27 15:16:48.475924] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:12:20.495 [2024-11-27 15:16:48.475937] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:20.495 [2024-11-27 15:16:48.478305] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:20.495 [2024-11-27 15:16:48.478407] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:20.495 spare 00:12:20.495 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:20.495 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:12:20.495 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:20.495 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:20.495 [2024-11-27 15:16:48.487881] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:20.495 [2024-11-27 15:16:48.489999] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:20.495 [2024-11-27 15:16:48.490077] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:20.495 [2024-11-27 15:16:48.490124] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:20.495 [2024-11-27 15:16:48.490208] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:12:20.495 [2024-11-27 15:16:48.490226] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:12:20.495 [2024-11-27 15:16:48.490505] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:12:20.495 [2024-11-27 15:16:48.490658] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:12:20.495 [2024-11-27 15:16:48.490673] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:12:20.495 [2024-11-27 15:16:48.490795] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:20.495 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:20.495 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:12:20.495 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:20.495 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:20.495 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:20.495 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:20.495 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:20.495 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:20.495 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:20.495 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:20.495 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:20.495 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:20.495 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:20.495 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:20.495 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:20.495 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:20.495 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:20.495 "name": "raid_bdev1", 00:12:20.495 "uuid": "844da588-a5c4-4c5f-8d81-60504a483e55", 00:12:20.495 "strip_size_kb": 0, 00:12:20.495 "state": "online", 00:12:20.495 "raid_level": "raid1", 00:12:20.495 "superblock": false, 00:12:20.495 "num_base_bdevs": 4, 00:12:20.495 "num_base_bdevs_discovered": 4, 00:12:20.495 "num_base_bdevs_operational": 4, 00:12:20.495 "base_bdevs_list": [ 00:12:20.495 { 00:12:20.495 "name": "BaseBdev1", 00:12:20.495 "uuid": "6b5337f3-7779-5666-b728-2ffbc84fdc51", 00:12:20.495 "is_configured": true, 00:12:20.495 "data_offset": 0, 00:12:20.495 "data_size": 65536 00:12:20.495 }, 00:12:20.495 { 00:12:20.495 "name": "BaseBdev2", 00:12:20.495 "uuid": "308930e9-8902-5f2d-aa2c-bbbe3cdd4820", 00:12:20.495 "is_configured": true, 00:12:20.495 "data_offset": 0, 00:12:20.495 "data_size": 65536 00:12:20.495 }, 00:12:20.495 { 00:12:20.495 "name": "BaseBdev3", 00:12:20.495 "uuid": "248ffc38-d706-5d97-8411-1ff90fa388d7", 00:12:20.495 "is_configured": true, 00:12:20.495 "data_offset": 0, 00:12:20.495 "data_size": 65536 00:12:20.495 }, 00:12:20.495 { 00:12:20.495 "name": "BaseBdev4", 00:12:20.495 "uuid": "d339b95f-fcab-585b-92d9-9548bb936e85", 00:12:20.495 "is_configured": true, 00:12:20.495 "data_offset": 0, 00:12:20.495 "data_size": 65536 00:12:20.495 } 00:12:20.495 ] 00:12:20.495 }' 00:12:20.495 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:20.495 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:21.065 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:12:21.065 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:21.065 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:21.065 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:21.065 [2024-11-27 15:16:48.919583] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:21.065 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:21.065 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:12:21.065 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:21.065 15:16:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:12:21.065 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:21.065 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:21.065 15:16:48 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:21.065 15:16:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:12:21.065 15:16:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:12:21.065 15:16:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:12:21.065 15:16:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:12:21.065 15:16:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:12:21.065 15:16:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:21.065 15:16:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:12:21.065 15:16:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:21.065 15:16:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:12:21.065 15:16:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:21.065 15:16:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:12:21.065 15:16:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:21.065 15:16:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:21.065 15:16:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:12:21.323 [2024-11-27 15:16:49.194870] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:12:21.323 /dev/nbd0 00:12:21.323 15:16:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:21.323 15:16:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:21.323 15:16:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:12:21.323 15:16:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:12:21.323 15:16:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:21.323 15:16:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:21.323 15:16:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:12:21.323 15:16:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:12:21.323 15:16:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:21.323 15:16:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:21.323 15:16:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:21.323 1+0 records in 00:12:21.323 1+0 records out 00:12:21.323 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00038247 s, 10.7 MB/s 00:12:21.323 15:16:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:21.323 15:16:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:12:21.323 15:16:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:21.323 15:16:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:21.323 15:16:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:12:21.323 15:16:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:21.323 15:16:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:21.323 15:16:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:12:21.323 15:16:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:12:21.323 15:16:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=65536 oflag=direct 00:12:27.886 65536+0 records in 00:12:27.886 65536+0 records out 00:12:27.886 33554432 bytes (34 MB, 32 MiB) copied, 6.34336 s, 5.3 MB/s 00:12:27.886 15:16:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:12:27.886 15:16:55 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:27.886 15:16:55 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:12:27.886 15:16:55 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:27.886 15:16:55 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:12:27.886 15:16:55 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:27.886 15:16:55 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:27.886 [2024-11-27 15:16:55.798964] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:27.886 15:16:55 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:27.886 15:16:55 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:27.886 15:16:55 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:27.886 15:16:55 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:27.886 15:16:55 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:27.886 15:16:55 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:27.886 15:16:55 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:12:27.886 15:16:55 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:12:27.886 15:16:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:12:27.886 15:16:55 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:27.886 15:16:55 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:27.886 [2024-11-27 15:16:55.830979] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:27.886 15:16:55 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:27.886 15:16:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:27.886 15:16:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:27.886 15:16:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:27.886 15:16:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:27.887 15:16:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:27.887 15:16:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:27.887 15:16:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:27.887 15:16:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:27.887 15:16:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:27.887 15:16:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:27.887 15:16:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:27.887 15:16:55 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:27.887 15:16:55 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:27.887 15:16:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:27.887 15:16:55 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:27.887 15:16:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:27.887 "name": "raid_bdev1", 00:12:27.887 "uuid": "844da588-a5c4-4c5f-8d81-60504a483e55", 00:12:27.887 "strip_size_kb": 0, 00:12:27.887 "state": "online", 00:12:27.887 "raid_level": "raid1", 00:12:27.887 "superblock": false, 00:12:27.887 "num_base_bdevs": 4, 00:12:27.887 "num_base_bdevs_discovered": 3, 00:12:27.887 "num_base_bdevs_operational": 3, 00:12:27.887 "base_bdevs_list": [ 00:12:27.887 { 00:12:27.887 "name": null, 00:12:27.887 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:27.887 "is_configured": false, 00:12:27.887 "data_offset": 0, 00:12:27.887 "data_size": 65536 00:12:27.887 }, 00:12:27.887 { 00:12:27.887 "name": "BaseBdev2", 00:12:27.887 "uuid": "308930e9-8902-5f2d-aa2c-bbbe3cdd4820", 00:12:27.887 "is_configured": true, 00:12:27.887 "data_offset": 0, 00:12:27.887 "data_size": 65536 00:12:27.887 }, 00:12:27.887 { 00:12:27.887 "name": "BaseBdev3", 00:12:27.887 "uuid": "248ffc38-d706-5d97-8411-1ff90fa388d7", 00:12:27.887 "is_configured": true, 00:12:27.887 "data_offset": 0, 00:12:27.887 "data_size": 65536 00:12:27.887 }, 00:12:27.887 { 00:12:27.887 "name": "BaseBdev4", 00:12:27.887 "uuid": "d339b95f-fcab-585b-92d9-9548bb936e85", 00:12:27.887 "is_configured": true, 00:12:27.887 "data_offset": 0, 00:12:27.887 "data_size": 65536 00:12:27.887 } 00:12:27.887 ] 00:12:27.887 }' 00:12:27.887 15:16:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:27.887 15:16:55 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:28.144 15:16:56 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:28.144 15:16:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:28.144 15:16:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:28.144 [2024-11-27 15:16:56.218415] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:28.144 [2024-11-27 15:16:56.226055] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d09bd0 00:12:28.144 15:16:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:28.144 15:16:56 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:12:28.144 [2024-11-27 15:16:56.228582] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:29.522 "name": "raid_bdev1", 00:12:29.522 "uuid": "844da588-a5c4-4c5f-8d81-60504a483e55", 00:12:29.522 "strip_size_kb": 0, 00:12:29.522 "state": "online", 00:12:29.522 "raid_level": "raid1", 00:12:29.522 "superblock": false, 00:12:29.522 "num_base_bdevs": 4, 00:12:29.522 "num_base_bdevs_discovered": 4, 00:12:29.522 "num_base_bdevs_operational": 4, 00:12:29.522 "process": { 00:12:29.522 "type": "rebuild", 00:12:29.522 "target": "spare", 00:12:29.522 "progress": { 00:12:29.522 "blocks": 20480, 00:12:29.522 "percent": 31 00:12:29.522 } 00:12:29.522 }, 00:12:29.522 "base_bdevs_list": [ 00:12:29.522 { 00:12:29.522 "name": "spare", 00:12:29.522 "uuid": "8e198c6b-46aa-573b-84c4-4ee156d701dc", 00:12:29.522 "is_configured": true, 00:12:29.522 "data_offset": 0, 00:12:29.522 "data_size": 65536 00:12:29.522 }, 00:12:29.522 { 00:12:29.522 "name": "BaseBdev2", 00:12:29.522 "uuid": "308930e9-8902-5f2d-aa2c-bbbe3cdd4820", 00:12:29.522 "is_configured": true, 00:12:29.522 "data_offset": 0, 00:12:29.522 "data_size": 65536 00:12:29.522 }, 00:12:29.522 { 00:12:29.522 "name": "BaseBdev3", 00:12:29.522 "uuid": "248ffc38-d706-5d97-8411-1ff90fa388d7", 00:12:29.522 "is_configured": true, 00:12:29.522 "data_offset": 0, 00:12:29.522 "data_size": 65536 00:12:29.522 }, 00:12:29.522 { 00:12:29.522 "name": "BaseBdev4", 00:12:29.522 "uuid": "d339b95f-fcab-585b-92d9-9548bb936e85", 00:12:29.522 "is_configured": true, 00:12:29.522 "data_offset": 0, 00:12:29.522 "data_size": 65536 00:12:29.522 } 00:12:29.522 ] 00:12:29.522 }' 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:29.522 [2024-11-27 15:16:57.388667] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:29.522 [2024-11-27 15:16:57.438515] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:29.522 [2024-11-27 15:16:57.439183] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:29.522 [2024-11-27 15:16:57.439224] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:29.522 [2024-11-27 15:16:57.439239] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:29.522 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:29.522 "name": "raid_bdev1", 00:12:29.522 "uuid": "844da588-a5c4-4c5f-8d81-60504a483e55", 00:12:29.522 "strip_size_kb": 0, 00:12:29.522 "state": "online", 00:12:29.522 "raid_level": "raid1", 00:12:29.522 "superblock": false, 00:12:29.522 "num_base_bdevs": 4, 00:12:29.522 "num_base_bdevs_discovered": 3, 00:12:29.522 "num_base_bdevs_operational": 3, 00:12:29.522 "base_bdevs_list": [ 00:12:29.522 { 00:12:29.522 "name": null, 00:12:29.522 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:29.522 "is_configured": false, 00:12:29.522 "data_offset": 0, 00:12:29.522 "data_size": 65536 00:12:29.522 }, 00:12:29.522 { 00:12:29.522 "name": "BaseBdev2", 00:12:29.522 "uuid": "308930e9-8902-5f2d-aa2c-bbbe3cdd4820", 00:12:29.522 "is_configured": true, 00:12:29.522 "data_offset": 0, 00:12:29.522 "data_size": 65536 00:12:29.522 }, 00:12:29.522 { 00:12:29.522 "name": "BaseBdev3", 00:12:29.522 "uuid": "248ffc38-d706-5d97-8411-1ff90fa388d7", 00:12:29.522 "is_configured": true, 00:12:29.522 "data_offset": 0, 00:12:29.522 "data_size": 65536 00:12:29.522 }, 00:12:29.522 { 00:12:29.522 "name": "BaseBdev4", 00:12:29.522 "uuid": "d339b95f-fcab-585b-92d9-9548bb936e85", 00:12:29.522 "is_configured": true, 00:12:29.522 "data_offset": 0, 00:12:29.522 "data_size": 65536 00:12:29.522 } 00:12:29.522 ] 00:12:29.522 }' 00:12:29.523 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:29.523 15:16:57 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:29.780 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:29.781 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:29.781 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:29.781 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:29.781 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:29.781 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:29.781 15:16:57 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:29.781 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:30.040 15:16:57 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:30.040 15:16:57 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:30.040 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:30.040 "name": "raid_bdev1", 00:12:30.040 "uuid": "844da588-a5c4-4c5f-8d81-60504a483e55", 00:12:30.040 "strip_size_kb": 0, 00:12:30.040 "state": "online", 00:12:30.040 "raid_level": "raid1", 00:12:30.040 "superblock": false, 00:12:30.040 "num_base_bdevs": 4, 00:12:30.040 "num_base_bdevs_discovered": 3, 00:12:30.040 "num_base_bdevs_operational": 3, 00:12:30.040 "base_bdevs_list": [ 00:12:30.040 { 00:12:30.040 "name": null, 00:12:30.040 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:30.040 "is_configured": false, 00:12:30.040 "data_offset": 0, 00:12:30.040 "data_size": 65536 00:12:30.040 }, 00:12:30.040 { 00:12:30.040 "name": "BaseBdev2", 00:12:30.040 "uuid": "308930e9-8902-5f2d-aa2c-bbbe3cdd4820", 00:12:30.040 "is_configured": true, 00:12:30.040 "data_offset": 0, 00:12:30.040 "data_size": 65536 00:12:30.040 }, 00:12:30.040 { 00:12:30.040 "name": "BaseBdev3", 00:12:30.040 "uuid": "248ffc38-d706-5d97-8411-1ff90fa388d7", 00:12:30.040 "is_configured": true, 00:12:30.040 "data_offset": 0, 00:12:30.040 "data_size": 65536 00:12:30.040 }, 00:12:30.040 { 00:12:30.040 "name": "BaseBdev4", 00:12:30.040 "uuid": "d339b95f-fcab-585b-92d9-9548bb936e85", 00:12:30.040 "is_configured": true, 00:12:30.040 "data_offset": 0, 00:12:30.040 "data_size": 65536 00:12:30.040 } 00:12:30.040 ] 00:12:30.040 }' 00:12:30.040 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:30.040 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:30.040 15:16:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:30.040 15:16:58 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:30.040 15:16:58 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:30.040 15:16:58 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:30.040 15:16:58 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:30.040 [2024-11-27 15:16:58.014724] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:30.040 [2024-11-27 15:16:58.022287] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d09ca0 00:12:30.040 15:16:58 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:30.040 15:16:58 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:12:30.040 [2024-11-27 15:16:58.024753] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:30.977 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:30.977 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:30.977 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:30.977 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:30.977 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:30.977 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:30.977 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:30.977 15:16:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:30.977 15:16:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:30.977 15:16:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:30.977 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:30.977 "name": "raid_bdev1", 00:12:30.977 "uuid": "844da588-a5c4-4c5f-8d81-60504a483e55", 00:12:30.977 "strip_size_kb": 0, 00:12:30.977 "state": "online", 00:12:30.977 "raid_level": "raid1", 00:12:30.977 "superblock": false, 00:12:30.977 "num_base_bdevs": 4, 00:12:30.977 "num_base_bdevs_discovered": 4, 00:12:30.977 "num_base_bdevs_operational": 4, 00:12:30.977 "process": { 00:12:30.977 "type": "rebuild", 00:12:30.977 "target": "spare", 00:12:30.977 "progress": { 00:12:30.977 "blocks": 20480, 00:12:30.977 "percent": 31 00:12:30.977 } 00:12:30.977 }, 00:12:30.977 "base_bdevs_list": [ 00:12:30.977 { 00:12:30.977 "name": "spare", 00:12:30.977 "uuid": "8e198c6b-46aa-573b-84c4-4ee156d701dc", 00:12:30.977 "is_configured": true, 00:12:30.977 "data_offset": 0, 00:12:30.977 "data_size": 65536 00:12:30.977 }, 00:12:30.977 { 00:12:30.977 "name": "BaseBdev2", 00:12:30.977 "uuid": "308930e9-8902-5f2d-aa2c-bbbe3cdd4820", 00:12:30.977 "is_configured": true, 00:12:30.977 "data_offset": 0, 00:12:30.977 "data_size": 65536 00:12:30.977 }, 00:12:30.977 { 00:12:30.977 "name": "BaseBdev3", 00:12:30.977 "uuid": "248ffc38-d706-5d97-8411-1ff90fa388d7", 00:12:30.977 "is_configured": true, 00:12:30.977 "data_offset": 0, 00:12:30.977 "data_size": 65536 00:12:30.977 }, 00:12:30.977 { 00:12:30.977 "name": "BaseBdev4", 00:12:30.977 "uuid": "d339b95f-fcab-585b-92d9-9548bb936e85", 00:12:30.977 "is_configured": true, 00:12:30.977 "data_offset": 0, 00:12:30.977 "data_size": 65536 00:12:30.977 } 00:12:30.977 ] 00:12:30.977 }' 00:12:30.977 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:31.235 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:31.235 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:31.235 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:31.235 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:12:31.235 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:12:31.235 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:12:31.235 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:12:31.235 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:12:31.235 15:16:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.235 15:16:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:31.235 [2024-11-27 15:16:59.180443] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:31.235 [2024-11-27 15:16:59.233914] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000d09ca0 00:12:31.235 15:16:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:31.235 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:12:31.235 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:12:31.235 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:31.235 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:31.235 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:31.235 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:31.235 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:31.235 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:31.235 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:31.235 15:16:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.235 15:16:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:31.235 15:16:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:31.235 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:31.235 "name": "raid_bdev1", 00:12:31.236 "uuid": "844da588-a5c4-4c5f-8d81-60504a483e55", 00:12:31.236 "strip_size_kb": 0, 00:12:31.236 "state": "online", 00:12:31.236 "raid_level": "raid1", 00:12:31.236 "superblock": false, 00:12:31.236 "num_base_bdevs": 4, 00:12:31.236 "num_base_bdevs_discovered": 3, 00:12:31.236 "num_base_bdevs_operational": 3, 00:12:31.236 "process": { 00:12:31.236 "type": "rebuild", 00:12:31.236 "target": "spare", 00:12:31.236 "progress": { 00:12:31.236 "blocks": 24576, 00:12:31.236 "percent": 37 00:12:31.236 } 00:12:31.236 }, 00:12:31.236 "base_bdevs_list": [ 00:12:31.236 { 00:12:31.236 "name": "spare", 00:12:31.236 "uuid": "8e198c6b-46aa-573b-84c4-4ee156d701dc", 00:12:31.236 "is_configured": true, 00:12:31.236 "data_offset": 0, 00:12:31.236 "data_size": 65536 00:12:31.236 }, 00:12:31.236 { 00:12:31.236 "name": null, 00:12:31.236 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:31.236 "is_configured": false, 00:12:31.236 "data_offset": 0, 00:12:31.236 "data_size": 65536 00:12:31.236 }, 00:12:31.236 { 00:12:31.236 "name": "BaseBdev3", 00:12:31.236 "uuid": "248ffc38-d706-5d97-8411-1ff90fa388d7", 00:12:31.236 "is_configured": true, 00:12:31.236 "data_offset": 0, 00:12:31.236 "data_size": 65536 00:12:31.236 }, 00:12:31.236 { 00:12:31.236 "name": "BaseBdev4", 00:12:31.236 "uuid": "d339b95f-fcab-585b-92d9-9548bb936e85", 00:12:31.236 "is_configured": true, 00:12:31.236 "data_offset": 0, 00:12:31.236 "data_size": 65536 00:12:31.236 } 00:12:31.236 ] 00:12:31.236 }' 00:12:31.236 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:31.236 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:31.495 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:31.495 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:31.495 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=360 00:12:31.495 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:31.495 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:31.495 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:31.495 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:31.495 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:31.495 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:31.495 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:31.495 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:31.495 15:16:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.495 15:16:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:31.495 15:16:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:31.495 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:31.495 "name": "raid_bdev1", 00:12:31.495 "uuid": "844da588-a5c4-4c5f-8d81-60504a483e55", 00:12:31.495 "strip_size_kb": 0, 00:12:31.495 "state": "online", 00:12:31.495 "raid_level": "raid1", 00:12:31.495 "superblock": false, 00:12:31.495 "num_base_bdevs": 4, 00:12:31.495 "num_base_bdevs_discovered": 3, 00:12:31.495 "num_base_bdevs_operational": 3, 00:12:31.495 "process": { 00:12:31.495 "type": "rebuild", 00:12:31.495 "target": "spare", 00:12:31.495 "progress": { 00:12:31.495 "blocks": 26624, 00:12:31.495 "percent": 40 00:12:31.495 } 00:12:31.495 }, 00:12:31.495 "base_bdevs_list": [ 00:12:31.495 { 00:12:31.495 "name": "spare", 00:12:31.495 "uuid": "8e198c6b-46aa-573b-84c4-4ee156d701dc", 00:12:31.495 "is_configured": true, 00:12:31.495 "data_offset": 0, 00:12:31.495 "data_size": 65536 00:12:31.495 }, 00:12:31.495 { 00:12:31.495 "name": null, 00:12:31.495 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:31.495 "is_configured": false, 00:12:31.495 "data_offset": 0, 00:12:31.495 "data_size": 65536 00:12:31.495 }, 00:12:31.495 { 00:12:31.495 "name": "BaseBdev3", 00:12:31.495 "uuid": "248ffc38-d706-5d97-8411-1ff90fa388d7", 00:12:31.495 "is_configured": true, 00:12:31.495 "data_offset": 0, 00:12:31.495 "data_size": 65536 00:12:31.495 }, 00:12:31.495 { 00:12:31.495 "name": "BaseBdev4", 00:12:31.495 "uuid": "d339b95f-fcab-585b-92d9-9548bb936e85", 00:12:31.495 "is_configured": true, 00:12:31.495 "data_offset": 0, 00:12:31.495 "data_size": 65536 00:12:31.495 } 00:12:31.495 ] 00:12:31.495 }' 00:12:31.495 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:31.496 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:31.496 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:31.496 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:31.496 15:16:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:32.433 15:17:00 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:32.433 15:17:00 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:32.433 15:17:00 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:32.433 15:17:00 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:32.433 15:17:00 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:32.433 15:17:00 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:32.433 15:17:00 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:32.433 15:17:00 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:32.433 15:17:00 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:32.433 15:17:00 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:32.433 15:17:00 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:32.691 15:17:00 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:32.691 "name": "raid_bdev1", 00:12:32.691 "uuid": "844da588-a5c4-4c5f-8d81-60504a483e55", 00:12:32.691 "strip_size_kb": 0, 00:12:32.691 "state": "online", 00:12:32.691 "raid_level": "raid1", 00:12:32.691 "superblock": false, 00:12:32.691 "num_base_bdevs": 4, 00:12:32.691 "num_base_bdevs_discovered": 3, 00:12:32.691 "num_base_bdevs_operational": 3, 00:12:32.691 "process": { 00:12:32.691 "type": "rebuild", 00:12:32.691 "target": "spare", 00:12:32.691 "progress": { 00:12:32.691 "blocks": 49152, 00:12:32.691 "percent": 75 00:12:32.691 } 00:12:32.691 }, 00:12:32.691 "base_bdevs_list": [ 00:12:32.691 { 00:12:32.691 "name": "spare", 00:12:32.691 "uuid": "8e198c6b-46aa-573b-84c4-4ee156d701dc", 00:12:32.691 "is_configured": true, 00:12:32.691 "data_offset": 0, 00:12:32.691 "data_size": 65536 00:12:32.691 }, 00:12:32.691 { 00:12:32.691 "name": null, 00:12:32.691 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:32.691 "is_configured": false, 00:12:32.691 "data_offset": 0, 00:12:32.691 "data_size": 65536 00:12:32.691 }, 00:12:32.691 { 00:12:32.691 "name": "BaseBdev3", 00:12:32.691 "uuid": "248ffc38-d706-5d97-8411-1ff90fa388d7", 00:12:32.691 "is_configured": true, 00:12:32.691 "data_offset": 0, 00:12:32.691 "data_size": 65536 00:12:32.691 }, 00:12:32.691 { 00:12:32.691 "name": "BaseBdev4", 00:12:32.691 "uuid": "d339b95f-fcab-585b-92d9-9548bb936e85", 00:12:32.691 "is_configured": true, 00:12:32.691 "data_offset": 0, 00:12:32.691 "data_size": 65536 00:12:32.691 } 00:12:32.691 ] 00:12:32.691 }' 00:12:32.691 15:17:00 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:32.691 15:17:00 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:32.691 15:17:00 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:32.691 15:17:00 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:32.691 15:17:00 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:33.261 [2024-11-27 15:17:01.249025] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:12:33.261 [2024-11-27 15:17:01.249225] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:12:33.261 [2024-11-27 15:17:01.249896] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:33.831 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:33.831 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:33.831 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:33.831 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:33.831 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:33.831 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:33.831 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:33.831 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:33.831 15:17:01 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:33.831 15:17:01 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:33.831 15:17:01 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:33.831 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:33.831 "name": "raid_bdev1", 00:12:33.831 "uuid": "844da588-a5c4-4c5f-8d81-60504a483e55", 00:12:33.831 "strip_size_kb": 0, 00:12:33.831 "state": "online", 00:12:33.831 "raid_level": "raid1", 00:12:33.831 "superblock": false, 00:12:33.831 "num_base_bdevs": 4, 00:12:33.831 "num_base_bdevs_discovered": 3, 00:12:33.831 "num_base_bdevs_operational": 3, 00:12:33.831 "base_bdevs_list": [ 00:12:33.831 { 00:12:33.831 "name": "spare", 00:12:33.831 "uuid": "8e198c6b-46aa-573b-84c4-4ee156d701dc", 00:12:33.831 "is_configured": true, 00:12:33.831 "data_offset": 0, 00:12:33.831 "data_size": 65536 00:12:33.831 }, 00:12:33.831 { 00:12:33.831 "name": null, 00:12:33.831 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:33.831 "is_configured": false, 00:12:33.831 "data_offset": 0, 00:12:33.831 "data_size": 65536 00:12:33.831 }, 00:12:33.831 { 00:12:33.831 "name": "BaseBdev3", 00:12:33.831 "uuid": "248ffc38-d706-5d97-8411-1ff90fa388d7", 00:12:33.831 "is_configured": true, 00:12:33.831 "data_offset": 0, 00:12:33.831 "data_size": 65536 00:12:33.831 }, 00:12:33.831 { 00:12:33.831 "name": "BaseBdev4", 00:12:33.831 "uuid": "d339b95f-fcab-585b-92d9-9548bb936e85", 00:12:33.831 "is_configured": true, 00:12:33.831 "data_offset": 0, 00:12:33.831 "data_size": 65536 00:12:33.831 } 00:12:33.831 ] 00:12:33.831 }' 00:12:33.831 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:33.831 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:12:33.831 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:33.831 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:12:33.831 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:12:33.831 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:33.831 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:33.831 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:33.831 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:33.831 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:33.831 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:33.831 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:33.831 15:17:01 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:33.831 15:17:01 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:33.831 15:17:01 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:33.831 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:33.831 "name": "raid_bdev1", 00:12:33.831 "uuid": "844da588-a5c4-4c5f-8d81-60504a483e55", 00:12:33.832 "strip_size_kb": 0, 00:12:33.832 "state": "online", 00:12:33.832 "raid_level": "raid1", 00:12:33.832 "superblock": false, 00:12:33.832 "num_base_bdevs": 4, 00:12:33.832 "num_base_bdevs_discovered": 3, 00:12:33.832 "num_base_bdevs_operational": 3, 00:12:33.832 "base_bdevs_list": [ 00:12:33.832 { 00:12:33.832 "name": "spare", 00:12:33.832 "uuid": "8e198c6b-46aa-573b-84c4-4ee156d701dc", 00:12:33.832 "is_configured": true, 00:12:33.832 "data_offset": 0, 00:12:33.832 "data_size": 65536 00:12:33.832 }, 00:12:33.832 { 00:12:33.832 "name": null, 00:12:33.832 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:33.832 "is_configured": false, 00:12:33.832 "data_offset": 0, 00:12:33.832 "data_size": 65536 00:12:33.832 }, 00:12:33.832 { 00:12:33.832 "name": "BaseBdev3", 00:12:33.832 "uuid": "248ffc38-d706-5d97-8411-1ff90fa388d7", 00:12:33.832 "is_configured": true, 00:12:33.832 "data_offset": 0, 00:12:33.832 "data_size": 65536 00:12:33.832 }, 00:12:33.832 { 00:12:33.832 "name": "BaseBdev4", 00:12:33.832 "uuid": "d339b95f-fcab-585b-92d9-9548bb936e85", 00:12:33.832 "is_configured": true, 00:12:33.832 "data_offset": 0, 00:12:33.832 "data_size": 65536 00:12:33.832 } 00:12:33.832 ] 00:12:33.832 }' 00:12:33.832 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:33.832 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:33.832 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:33.832 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:33.832 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:33.832 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:33.832 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:33.832 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:33.832 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:33.832 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:33.832 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:33.832 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:33.832 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:33.832 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:33.832 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:33.832 15:17:01 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:33.832 15:17:01 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:33.832 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:34.090 15:17:01 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:34.090 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:34.090 "name": "raid_bdev1", 00:12:34.090 "uuid": "844da588-a5c4-4c5f-8d81-60504a483e55", 00:12:34.090 "strip_size_kb": 0, 00:12:34.090 "state": "online", 00:12:34.090 "raid_level": "raid1", 00:12:34.090 "superblock": false, 00:12:34.090 "num_base_bdevs": 4, 00:12:34.090 "num_base_bdevs_discovered": 3, 00:12:34.090 "num_base_bdevs_operational": 3, 00:12:34.090 "base_bdevs_list": [ 00:12:34.090 { 00:12:34.090 "name": "spare", 00:12:34.090 "uuid": "8e198c6b-46aa-573b-84c4-4ee156d701dc", 00:12:34.090 "is_configured": true, 00:12:34.090 "data_offset": 0, 00:12:34.090 "data_size": 65536 00:12:34.090 }, 00:12:34.090 { 00:12:34.091 "name": null, 00:12:34.091 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:34.091 "is_configured": false, 00:12:34.091 "data_offset": 0, 00:12:34.091 "data_size": 65536 00:12:34.091 }, 00:12:34.091 { 00:12:34.091 "name": "BaseBdev3", 00:12:34.091 "uuid": "248ffc38-d706-5d97-8411-1ff90fa388d7", 00:12:34.091 "is_configured": true, 00:12:34.091 "data_offset": 0, 00:12:34.091 "data_size": 65536 00:12:34.091 }, 00:12:34.091 { 00:12:34.091 "name": "BaseBdev4", 00:12:34.091 "uuid": "d339b95f-fcab-585b-92d9-9548bb936e85", 00:12:34.091 "is_configured": true, 00:12:34.091 "data_offset": 0, 00:12:34.091 "data_size": 65536 00:12:34.091 } 00:12:34.091 ] 00:12:34.091 }' 00:12:34.091 15:17:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:34.091 15:17:01 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:34.350 15:17:02 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:34.350 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:34.350 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:34.350 [2024-11-27 15:17:02.359938] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:34.350 [2024-11-27 15:17:02.359979] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:34.350 [2024-11-27 15:17:02.360118] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:34.350 [2024-11-27 15:17:02.360227] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:34.350 [2024-11-27 15:17:02.360245] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:12:34.350 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:34.350 15:17:02 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:34.350 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:34.350 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:34.350 15:17:02 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:12:34.350 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:34.350 15:17:02 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:12:34.350 15:17:02 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:12:34.350 15:17:02 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:12:34.350 15:17:02 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:12:34.350 15:17:02 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:34.350 15:17:02 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:12:34.350 15:17:02 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:34.350 15:17:02 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:34.350 15:17:02 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:34.350 15:17:02 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:12:34.350 15:17:02 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:34.350 15:17:02 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:34.350 15:17:02 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:12:34.610 /dev/nbd0 00:12:34.610 15:17:02 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:34.610 15:17:02 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:34.610 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:12:34.610 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:12:34.610 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:34.610 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:34.610 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:12:34.610 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:12:34.610 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:34.610 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:34.610 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:34.610 1+0 records in 00:12:34.610 1+0 records out 00:12:34.610 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000368481 s, 11.1 MB/s 00:12:34.610 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:34.610 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:12:34.610 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:34.610 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:34.610 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:12:34.610 15:17:02 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:34.610 15:17:02 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:34.610 15:17:02 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:12:34.869 /dev/nbd1 00:12:34.869 15:17:02 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:34.869 15:17:02 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:34.869 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:12:34.869 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:12:34.869 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:34.869 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:34.869 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:12:34.869 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:12:34.869 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:34.869 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:34.869 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:34.869 1+0 records in 00:12:34.869 1+0 records out 00:12:34.869 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000259421 s, 15.8 MB/s 00:12:34.869 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:34.869 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:12:34.869 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:34.869 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:34.869 15:17:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:12:34.869 15:17:02 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:34.869 15:17:02 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:34.869 15:17:02 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:12:34.869 15:17:02 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:12:34.869 15:17:02 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:34.869 15:17:02 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:34.869 15:17:02 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:34.869 15:17:02 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:12:34.869 15:17:02 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:34.869 15:17:02 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:35.128 15:17:03 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:35.128 15:17:03 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:35.128 15:17:03 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:35.128 15:17:03 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:35.128 15:17:03 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:35.128 15:17:03 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:35.128 15:17:03 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:12:35.128 15:17:03 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:12:35.128 15:17:03 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:35.128 15:17:03 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:35.387 15:17:03 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:35.387 15:17:03 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:35.388 15:17:03 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:35.388 15:17:03 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:35.388 15:17:03 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:35.388 15:17:03 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:35.388 15:17:03 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:12:35.388 15:17:03 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:12:35.388 15:17:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:12:35.388 15:17:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 88356 00:12:35.388 15:17:03 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@954 -- # '[' -z 88356 ']' 00:12:35.388 15:17:03 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@958 -- # kill -0 88356 00:12:35.388 15:17:03 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@959 -- # uname 00:12:35.388 15:17:03 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:35.388 15:17:03 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 88356 00:12:35.388 15:17:03 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:35.388 15:17:03 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:35.388 killing process with pid 88356 00:12:35.388 15:17:03 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 88356' 00:12:35.388 15:17:03 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@973 -- # kill 88356 00:12:35.388 Received shutdown signal, test time was about 60.000000 seconds 00:12:35.388 00:12:35.388 Latency(us) 00:12:35.388 [2024-11-27T15:17:03.495Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:35.388 [2024-11-27T15:17:03.495Z] =================================================================================================================== 00:12:35.388 [2024-11-27T15:17:03.495Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:12:35.388 [2024-11-27 15:17:03.429842] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:35.388 15:17:03 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@978 -- # wait 88356 00:12:35.388 [2024-11-27 15:17:03.480816] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:35.647 15:17:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:12:35.647 00:12:35.647 real 0m16.390s 00:12:35.647 user 0m17.577s 00:12:35.647 sys 0m3.466s 00:12:35.647 15:17:03 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:35.647 15:17:03 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:35.647 ************************************ 00:12:35.647 END TEST raid_rebuild_test 00:12:35.647 ************************************ 00:12:35.647 15:17:03 bdev_raid -- bdev/bdev_raid.sh@979 -- # run_test raid_rebuild_test_sb raid_rebuild_test raid1 4 true false true 00:12:35.647 15:17:03 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:12:35.647 15:17:03 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:35.647 15:17:03 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:35.647 ************************************ 00:12:35.647 START TEST raid_rebuild_test_sb 00:12:35.647 ************************************ 00:12:35.647 15:17:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 4 true false true 00:12:35.647 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:12:35.647 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=88793 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 88793 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@835 -- # '[' -z 88793 ']' 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:35.906 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:35.906 15:17:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:35.906 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:35.906 Zero copy mechanism will not be used. 00:12:35.906 [2024-11-27 15:17:03.845766] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:12:35.906 [2024-11-27 15:17:03.845887] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid88793 ] 00:12:36.166 [2024-11-27 15:17:04.016455] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:36.166 [2024-11-27 15:17:04.042501] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:36.166 [2024-11-27 15:17:04.085607] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:36.166 [2024-11-27 15:17:04.085652] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:36.735 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:36.735 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # return 0 00:12:36.735 15:17:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:36.735 15:17:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:36.735 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:36.735 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:36.735 BaseBdev1_malloc 00:12:36.735 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:36.735 15:17:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:36.736 [2024-11-27 15:17:04.701630] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:36.736 [2024-11-27 15:17:04.701692] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:36.736 [2024-11-27 15:17:04.701734] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:36.736 [2024-11-27 15:17:04.701747] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:36.736 [2024-11-27 15:17:04.703923] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:36.736 [2024-11-27 15:17:04.703959] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:36.736 BaseBdev1 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:36.736 BaseBdev2_malloc 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:36.736 [2024-11-27 15:17:04.730228] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:12:36.736 [2024-11-27 15:17:04.730336] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:36.736 [2024-11-27 15:17:04.730374] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:12:36.736 [2024-11-27 15:17:04.730401] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:36.736 [2024-11-27 15:17:04.732514] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:36.736 [2024-11-27 15:17:04.732583] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:36.736 BaseBdev2 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:36.736 BaseBdev3_malloc 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:36.736 [2024-11-27 15:17:04.758754] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:12:36.736 [2024-11-27 15:17:04.758858] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:36.736 [2024-11-27 15:17:04.758898] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:12:36.736 [2024-11-27 15:17:04.758948] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:36.736 [2024-11-27 15:17:04.761079] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:36.736 [2024-11-27 15:17:04.761148] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:12:36.736 BaseBdev3 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:36.736 BaseBdev4_malloc 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:36.736 [2024-11-27 15:17:04.795591] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:12:36.736 [2024-11-27 15:17:04.795704] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:36.736 [2024-11-27 15:17:04.795749] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:12:36.736 [2024-11-27 15:17:04.795788] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:36.736 [2024-11-27 15:17:04.797831] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:36.736 [2024-11-27 15:17:04.797910] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:12:36.736 BaseBdev4 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:36.736 spare_malloc 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:36.736 spare_delay 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:36.736 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:36.736 [2024-11-27 15:17:04.836216] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:36.736 [2024-11-27 15:17:04.836303] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:36.736 [2024-11-27 15:17:04.836340] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:12:36.736 [2024-11-27 15:17:04.836369] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:36.736 [2024-11-27 15:17:04.838541] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:36.736 [2024-11-27 15:17:04.838572] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:36.995 spare 00:12:36.995 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:36.995 15:17:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:12:36.995 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:36.995 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:36.995 [2024-11-27 15:17:04.848265] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:36.995 [2024-11-27 15:17:04.850123] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:36.995 [2024-11-27 15:17:04.850228] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:36.995 [2024-11-27 15:17:04.850288] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:36.995 [2024-11-27 15:17:04.850487] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:12:36.995 [2024-11-27 15:17:04.850534] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:36.995 [2024-11-27 15:17:04.850806] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:12:36.995 [2024-11-27 15:17:04.851014] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:12:36.995 [2024-11-27 15:17:04.851060] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:12:36.995 [2024-11-27 15:17:04.851223] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:36.995 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:36.995 15:17:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:12:36.995 15:17:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:36.995 15:17:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:36.995 15:17:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:36.995 15:17:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:36.995 15:17:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:36.995 15:17:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:36.995 15:17:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:36.995 15:17:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:36.996 15:17:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:36.996 15:17:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:36.996 15:17:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:36.996 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:36.996 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:36.996 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:36.996 15:17:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:36.996 "name": "raid_bdev1", 00:12:36.996 "uuid": "2fa6bfe2-c1e7-4aed-bdfe-8d3d58cd3706", 00:12:36.996 "strip_size_kb": 0, 00:12:36.996 "state": "online", 00:12:36.996 "raid_level": "raid1", 00:12:36.996 "superblock": true, 00:12:36.996 "num_base_bdevs": 4, 00:12:36.996 "num_base_bdevs_discovered": 4, 00:12:36.996 "num_base_bdevs_operational": 4, 00:12:36.996 "base_bdevs_list": [ 00:12:36.996 { 00:12:36.996 "name": "BaseBdev1", 00:12:36.996 "uuid": "a4e7e0d4-e3e8-562b-9ec2-c36ea2b0d38b", 00:12:36.996 "is_configured": true, 00:12:36.996 "data_offset": 2048, 00:12:36.996 "data_size": 63488 00:12:36.996 }, 00:12:36.996 { 00:12:36.996 "name": "BaseBdev2", 00:12:36.996 "uuid": "31ede09a-780a-5b5d-ae6f-b56b8ebf6ba4", 00:12:36.996 "is_configured": true, 00:12:36.996 "data_offset": 2048, 00:12:36.996 "data_size": 63488 00:12:36.996 }, 00:12:36.996 { 00:12:36.996 "name": "BaseBdev3", 00:12:36.996 "uuid": "3971710a-3a38-55c4-9cb6-8c93f5904fd5", 00:12:36.996 "is_configured": true, 00:12:36.996 "data_offset": 2048, 00:12:36.996 "data_size": 63488 00:12:36.996 }, 00:12:36.996 { 00:12:36.996 "name": "BaseBdev4", 00:12:36.996 "uuid": "71a325e4-1ec4-580c-ad64-3d2109daa198", 00:12:36.996 "is_configured": true, 00:12:36.996 "data_offset": 2048, 00:12:36.996 "data_size": 63488 00:12:36.996 } 00:12:36.996 ] 00:12:36.996 }' 00:12:36.996 15:17:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:36.996 15:17:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:37.254 15:17:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:12:37.254 15:17:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:37.254 15:17:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.254 15:17:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:37.254 [2024-11-27 15:17:05.291899] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:37.254 15:17:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.254 15:17:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:12:37.254 15:17:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:37.254 15:17:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.254 15:17:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:37.254 15:17:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:12:37.254 15:17:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.514 15:17:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:12:37.514 15:17:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:12:37.514 15:17:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:12:37.514 15:17:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:12:37.514 15:17:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:12:37.514 15:17:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:37.514 15:17:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:12:37.514 15:17:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:37.514 15:17:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:12:37.514 15:17:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:37.514 15:17:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:12:37.514 15:17:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:37.514 15:17:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:37.514 15:17:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:12:37.514 [2024-11-27 15:17:05.559166] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:12:37.514 /dev/nbd0 00:12:37.514 15:17:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:37.514 15:17:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:37.514 15:17:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:12:37.514 15:17:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:12:37.514 15:17:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:37.514 15:17:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:37.514 15:17:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:12:37.514 15:17:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:12:37.514 15:17:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:37.514 15:17:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:37.514 15:17:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:37.514 1+0 records in 00:12:37.514 1+0 records out 00:12:37.514 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000555094 s, 7.4 MB/s 00:12:37.514 15:17:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:37.514 15:17:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:12:37.794 15:17:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:37.795 15:17:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:37.795 15:17:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:12:37.795 15:17:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:37.795 15:17:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:37.795 15:17:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:12:37.795 15:17:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:12:37.795 15:17:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=63488 oflag=direct 00:12:43.118 63488+0 records in 00:12:43.118 63488+0 records out 00:12:43.118 32505856 bytes (33 MB, 31 MiB) copied, 5.49821 s, 5.9 MB/s 00:12:43.118 15:17:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:12:43.118 15:17:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:43.118 15:17:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:12:43.118 15:17:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:43.118 15:17:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:12:43.118 15:17:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:43.118 15:17:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:43.376 [2024-11-27 15:17:11.314187] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:43.376 15:17:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:43.376 15:17:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:43.376 15:17:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:43.376 15:17:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:43.376 15:17:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:43.376 15:17:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:43.376 15:17:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:12:43.376 15:17:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:12:43.377 15:17:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:12:43.377 15:17:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:43.377 15:17:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:43.377 [2024-11-27 15:17:11.350184] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:43.377 15:17:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:43.377 15:17:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:43.377 15:17:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:43.377 15:17:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:43.377 15:17:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:43.377 15:17:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:43.377 15:17:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:43.377 15:17:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:43.377 15:17:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:43.377 15:17:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:43.377 15:17:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:43.377 15:17:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:43.377 15:17:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:43.377 15:17:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:43.377 15:17:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:43.377 15:17:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:43.377 15:17:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:43.377 "name": "raid_bdev1", 00:12:43.377 "uuid": "2fa6bfe2-c1e7-4aed-bdfe-8d3d58cd3706", 00:12:43.377 "strip_size_kb": 0, 00:12:43.377 "state": "online", 00:12:43.377 "raid_level": "raid1", 00:12:43.377 "superblock": true, 00:12:43.377 "num_base_bdevs": 4, 00:12:43.377 "num_base_bdevs_discovered": 3, 00:12:43.377 "num_base_bdevs_operational": 3, 00:12:43.377 "base_bdevs_list": [ 00:12:43.377 { 00:12:43.377 "name": null, 00:12:43.377 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:43.377 "is_configured": false, 00:12:43.377 "data_offset": 0, 00:12:43.377 "data_size": 63488 00:12:43.377 }, 00:12:43.377 { 00:12:43.377 "name": "BaseBdev2", 00:12:43.377 "uuid": "31ede09a-780a-5b5d-ae6f-b56b8ebf6ba4", 00:12:43.377 "is_configured": true, 00:12:43.377 "data_offset": 2048, 00:12:43.377 "data_size": 63488 00:12:43.377 }, 00:12:43.377 { 00:12:43.377 "name": "BaseBdev3", 00:12:43.377 "uuid": "3971710a-3a38-55c4-9cb6-8c93f5904fd5", 00:12:43.377 "is_configured": true, 00:12:43.377 "data_offset": 2048, 00:12:43.377 "data_size": 63488 00:12:43.377 }, 00:12:43.377 { 00:12:43.377 "name": "BaseBdev4", 00:12:43.377 "uuid": "71a325e4-1ec4-580c-ad64-3d2109daa198", 00:12:43.377 "is_configured": true, 00:12:43.377 "data_offset": 2048, 00:12:43.377 "data_size": 63488 00:12:43.377 } 00:12:43.377 ] 00:12:43.377 }' 00:12:43.377 15:17:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:43.377 15:17:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:43.950 15:17:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:43.950 15:17:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:43.950 15:17:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:43.950 [2024-11-27 15:17:11.757514] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:43.950 [2024-11-27 15:17:11.761714] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000ca3360 00:12:43.950 15:17:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:43.950 [2024-11-27 15:17:11.763682] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:43.950 15:17:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:44.890 "name": "raid_bdev1", 00:12:44.890 "uuid": "2fa6bfe2-c1e7-4aed-bdfe-8d3d58cd3706", 00:12:44.890 "strip_size_kb": 0, 00:12:44.890 "state": "online", 00:12:44.890 "raid_level": "raid1", 00:12:44.890 "superblock": true, 00:12:44.890 "num_base_bdevs": 4, 00:12:44.890 "num_base_bdevs_discovered": 4, 00:12:44.890 "num_base_bdevs_operational": 4, 00:12:44.890 "process": { 00:12:44.890 "type": "rebuild", 00:12:44.890 "target": "spare", 00:12:44.890 "progress": { 00:12:44.890 "blocks": 20480, 00:12:44.890 "percent": 32 00:12:44.890 } 00:12:44.890 }, 00:12:44.890 "base_bdevs_list": [ 00:12:44.890 { 00:12:44.890 "name": "spare", 00:12:44.890 "uuid": "fd52ca22-1d55-53ff-b94e-f7e46e4b5f87", 00:12:44.890 "is_configured": true, 00:12:44.890 "data_offset": 2048, 00:12:44.890 "data_size": 63488 00:12:44.890 }, 00:12:44.890 { 00:12:44.890 "name": "BaseBdev2", 00:12:44.890 "uuid": "31ede09a-780a-5b5d-ae6f-b56b8ebf6ba4", 00:12:44.890 "is_configured": true, 00:12:44.890 "data_offset": 2048, 00:12:44.890 "data_size": 63488 00:12:44.890 }, 00:12:44.890 { 00:12:44.890 "name": "BaseBdev3", 00:12:44.890 "uuid": "3971710a-3a38-55c4-9cb6-8c93f5904fd5", 00:12:44.890 "is_configured": true, 00:12:44.890 "data_offset": 2048, 00:12:44.890 "data_size": 63488 00:12:44.890 }, 00:12:44.890 { 00:12:44.890 "name": "BaseBdev4", 00:12:44.890 "uuid": "71a325e4-1ec4-580c-ad64-3d2109daa198", 00:12:44.890 "is_configured": true, 00:12:44.890 "data_offset": 2048, 00:12:44.890 "data_size": 63488 00:12:44.890 } 00:12:44.890 ] 00:12:44.890 }' 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:44.890 [2024-11-27 15:17:12.916778] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:44.890 [2024-11-27 15:17:12.968696] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:44.890 [2024-11-27 15:17:12.968847] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:44.890 [2024-11-27 15:17:12.968907] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:44.890 [2024-11-27 15:17:12.968944] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:44.890 15:17:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:45.148 15:17:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:45.148 15:17:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:45.148 "name": "raid_bdev1", 00:12:45.148 "uuid": "2fa6bfe2-c1e7-4aed-bdfe-8d3d58cd3706", 00:12:45.148 "strip_size_kb": 0, 00:12:45.148 "state": "online", 00:12:45.148 "raid_level": "raid1", 00:12:45.148 "superblock": true, 00:12:45.148 "num_base_bdevs": 4, 00:12:45.148 "num_base_bdevs_discovered": 3, 00:12:45.148 "num_base_bdevs_operational": 3, 00:12:45.148 "base_bdevs_list": [ 00:12:45.148 { 00:12:45.148 "name": null, 00:12:45.148 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:45.148 "is_configured": false, 00:12:45.148 "data_offset": 0, 00:12:45.148 "data_size": 63488 00:12:45.148 }, 00:12:45.148 { 00:12:45.149 "name": "BaseBdev2", 00:12:45.149 "uuid": "31ede09a-780a-5b5d-ae6f-b56b8ebf6ba4", 00:12:45.149 "is_configured": true, 00:12:45.149 "data_offset": 2048, 00:12:45.149 "data_size": 63488 00:12:45.149 }, 00:12:45.149 { 00:12:45.149 "name": "BaseBdev3", 00:12:45.149 "uuid": "3971710a-3a38-55c4-9cb6-8c93f5904fd5", 00:12:45.149 "is_configured": true, 00:12:45.149 "data_offset": 2048, 00:12:45.149 "data_size": 63488 00:12:45.149 }, 00:12:45.149 { 00:12:45.149 "name": "BaseBdev4", 00:12:45.149 "uuid": "71a325e4-1ec4-580c-ad64-3d2109daa198", 00:12:45.149 "is_configured": true, 00:12:45.149 "data_offset": 2048, 00:12:45.149 "data_size": 63488 00:12:45.149 } 00:12:45.149 ] 00:12:45.149 }' 00:12:45.149 15:17:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:45.149 15:17:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:45.407 15:17:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:45.407 15:17:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:45.407 15:17:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:45.407 15:17:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:45.407 15:17:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:45.407 15:17:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:45.407 15:17:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:45.407 15:17:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:45.407 15:17:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:45.407 15:17:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:45.407 15:17:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:45.407 "name": "raid_bdev1", 00:12:45.407 "uuid": "2fa6bfe2-c1e7-4aed-bdfe-8d3d58cd3706", 00:12:45.407 "strip_size_kb": 0, 00:12:45.407 "state": "online", 00:12:45.407 "raid_level": "raid1", 00:12:45.407 "superblock": true, 00:12:45.407 "num_base_bdevs": 4, 00:12:45.407 "num_base_bdevs_discovered": 3, 00:12:45.407 "num_base_bdevs_operational": 3, 00:12:45.407 "base_bdevs_list": [ 00:12:45.407 { 00:12:45.407 "name": null, 00:12:45.407 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:45.407 "is_configured": false, 00:12:45.407 "data_offset": 0, 00:12:45.407 "data_size": 63488 00:12:45.407 }, 00:12:45.407 { 00:12:45.407 "name": "BaseBdev2", 00:12:45.407 "uuid": "31ede09a-780a-5b5d-ae6f-b56b8ebf6ba4", 00:12:45.407 "is_configured": true, 00:12:45.407 "data_offset": 2048, 00:12:45.407 "data_size": 63488 00:12:45.407 }, 00:12:45.407 { 00:12:45.407 "name": "BaseBdev3", 00:12:45.407 "uuid": "3971710a-3a38-55c4-9cb6-8c93f5904fd5", 00:12:45.407 "is_configured": true, 00:12:45.407 "data_offset": 2048, 00:12:45.407 "data_size": 63488 00:12:45.407 }, 00:12:45.407 { 00:12:45.407 "name": "BaseBdev4", 00:12:45.407 "uuid": "71a325e4-1ec4-580c-ad64-3d2109daa198", 00:12:45.407 "is_configured": true, 00:12:45.407 "data_offset": 2048, 00:12:45.407 "data_size": 63488 00:12:45.407 } 00:12:45.407 ] 00:12:45.407 }' 00:12:45.407 15:17:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:45.666 15:17:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:45.666 15:17:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:45.666 15:17:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:45.666 15:17:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:45.666 15:17:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:45.666 15:17:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:45.666 [2024-11-27 15:17:13.568386] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:45.666 [2024-11-27 15:17:13.572580] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000ca3430 00:12:45.666 15:17:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:45.666 15:17:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:12:45.666 [2024-11-27 15:17:13.574497] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:46.603 15:17:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:46.603 15:17:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:46.603 15:17:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:46.603 15:17:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:46.603 15:17:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:46.603 15:17:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:46.603 15:17:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:46.603 15:17:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:46.603 15:17:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:46.603 15:17:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:46.603 15:17:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:46.603 "name": "raid_bdev1", 00:12:46.603 "uuid": "2fa6bfe2-c1e7-4aed-bdfe-8d3d58cd3706", 00:12:46.603 "strip_size_kb": 0, 00:12:46.603 "state": "online", 00:12:46.603 "raid_level": "raid1", 00:12:46.603 "superblock": true, 00:12:46.603 "num_base_bdevs": 4, 00:12:46.603 "num_base_bdevs_discovered": 4, 00:12:46.603 "num_base_bdevs_operational": 4, 00:12:46.603 "process": { 00:12:46.603 "type": "rebuild", 00:12:46.603 "target": "spare", 00:12:46.603 "progress": { 00:12:46.603 "blocks": 20480, 00:12:46.603 "percent": 32 00:12:46.603 } 00:12:46.603 }, 00:12:46.603 "base_bdevs_list": [ 00:12:46.603 { 00:12:46.604 "name": "spare", 00:12:46.604 "uuid": "fd52ca22-1d55-53ff-b94e-f7e46e4b5f87", 00:12:46.604 "is_configured": true, 00:12:46.604 "data_offset": 2048, 00:12:46.604 "data_size": 63488 00:12:46.604 }, 00:12:46.604 { 00:12:46.604 "name": "BaseBdev2", 00:12:46.604 "uuid": "31ede09a-780a-5b5d-ae6f-b56b8ebf6ba4", 00:12:46.604 "is_configured": true, 00:12:46.604 "data_offset": 2048, 00:12:46.604 "data_size": 63488 00:12:46.604 }, 00:12:46.604 { 00:12:46.604 "name": "BaseBdev3", 00:12:46.604 "uuid": "3971710a-3a38-55c4-9cb6-8c93f5904fd5", 00:12:46.604 "is_configured": true, 00:12:46.604 "data_offset": 2048, 00:12:46.604 "data_size": 63488 00:12:46.604 }, 00:12:46.604 { 00:12:46.604 "name": "BaseBdev4", 00:12:46.604 "uuid": "71a325e4-1ec4-580c-ad64-3d2109daa198", 00:12:46.604 "is_configured": true, 00:12:46.604 "data_offset": 2048, 00:12:46.604 "data_size": 63488 00:12:46.604 } 00:12:46.604 ] 00:12:46.604 }' 00:12:46.604 15:17:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:46.604 15:17:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:46.604 15:17:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:46.863 15:17:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:46.863 15:17:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:12:46.863 15:17:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:12:46.863 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:12:46.863 15:17:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:12:46.863 15:17:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:12:46.863 15:17:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:12:46.863 15:17:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:12:46.863 15:17:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:46.863 15:17:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:46.863 [2024-11-27 15:17:14.743291] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:46.863 [2024-11-27 15:17:14.878524] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000ca3430 00:12:46.863 15:17:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:46.863 15:17:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:12:46.863 15:17:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:12:46.863 15:17:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:46.863 15:17:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:46.863 15:17:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:46.863 15:17:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:46.863 15:17:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:46.863 15:17:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:46.863 15:17:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:46.863 15:17:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:46.863 15:17:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:46.863 15:17:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:46.863 15:17:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:46.863 "name": "raid_bdev1", 00:12:46.863 "uuid": "2fa6bfe2-c1e7-4aed-bdfe-8d3d58cd3706", 00:12:46.863 "strip_size_kb": 0, 00:12:46.863 "state": "online", 00:12:46.863 "raid_level": "raid1", 00:12:46.863 "superblock": true, 00:12:46.863 "num_base_bdevs": 4, 00:12:46.863 "num_base_bdevs_discovered": 3, 00:12:46.863 "num_base_bdevs_operational": 3, 00:12:46.863 "process": { 00:12:46.863 "type": "rebuild", 00:12:46.863 "target": "spare", 00:12:46.863 "progress": { 00:12:46.863 "blocks": 24576, 00:12:46.863 "percent": 38 00:12:46.863 } 00:12:46.863 }, 00:12:46.863 "base_bdevs_list": [ 00:12:46.863 { 00:12:46.863 "name": "spare", 00:12:46.863 "uuid": "fd52ca22-1d55-53ff-b94e-f7e46e4b5f87", 00:12:46.863 "is_configured": true, 00:12:46.863 "data_offset": 2048, 00:12:46.863 "data_size": 63488 00:12:46.863 }, 00:12:46.863 { 00:12:46.863 "name": null, 00:12:46.863 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:46.863 "is_configured": false, 00:12:46.863 "data_offset": 0, 00:12:46.863 "data_size": 63488 00:12:46.863 }, 00:12:46.863 { 00:12:46.863 "name": "BaseBdev3", 00:12:46.863 "uuid": "3971710a-3a38-55c4-9cb6-8c93f5904fd5", 00:12:46.863 "is_configured": true, 00:12:46.863 "data_offset": 2048, 00:12:46.863 "data_size": 63488 00:12:46.863 }, 00:12:46.863 { 00:12:46.863 "name": "BaseBdev4", 00:12:46.864 "uuid": "71a325e4-1ec4-580c-ad64-3d2109daa198", 00:12:46.864 "is_configured": true, 00:12:46.864 "data_offset": 2048, 00:12:46.864 "data_size": 63488 00:12:46.864 } 00:12:46.864 ] 00:12:46.864 }' 00:12:46.864 15:17:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:47.123 15:17:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:47.123 15:17:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:47.123 15:17:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:47.123 15:17:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=376 00:12:47.123 15:17:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:47.123 15:17:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:47.123 15:17:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:47.123 15:17:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:47.123 15:17:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:47.123 15:17:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:47.123 15:17:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:47.123 15:17:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:47.123 15:17:15 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:47.123 15:17:15 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:47.123 15:17:15 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:47.123 15:17:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:47.123 "name": "raid_bdev1", 00:12:47.123 "uuid": "2fa6bfe2-c1e7-4aed-bdfe-8d3d58cd3706", 00:12:47.123 "strip_size_kb": 0, 00:12:47.123 "state": "online", 00:12:47.123 "raid_level": "raid1", 00:12:47.123 "superblock": true, 00:12:47.123 "num_base_bdevs": 4, 00:12:47.123 "num_base_bdevs_discovered": 3, 00:12:47.123 "num_base_bdevs_operational": 3, 00:12:47.123 "process": { 00:12:47.123 "type": "rebuild", 00:12:47.123 "target": "spare", 00:12:47.123 "progress": { 00:12:47.123 "blocks": 26624, 00:12:47.123 "percent": 41 00:12:47.123 } 00:12:47.123 }, 00:12:47.123 "base_bdevs_list": [ 00:12:47.123 { 00:12:47.123 "name": "spare", 00:12:47.123 "uuid": "fd52ca22-1d55-53ff-b94e-f7e46e4b5f87", 00:12:47.123 "is_configured": true, 00:12:47.123 "data_offset": 2048, 00:12:47.123 "data_size": 63488 00:12:47.123 }, 00:12:47.123 { 00:12:47.123 "name": null, 00:12:47.123 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:47.123 "is_configured": false, 00:12:47.123 "data_offset": 0, 00:12:47.123 "data_size": 63488 00:12:47.123 }, 00:12:47.123 { 00:12:47.123 "name": "BaseBdev3", 00:12:47.123 "uuid": "3971710a-3a38-55c4-9cb6-8c93f5904fd5", 00:12:47.123 "is_configured": true, 00:12:47.123 "data_offset": 2048, 00:12:47.123 "data_size": 63488 00:12:47.123 }, 00:12:47.123 { 00:12:47.123 "name": "BaseBdev4", 00:12:47.123 "uuid": "71a325e4-1ec4-580c-ad64-3d2109daa198", 00:12:47.123 "is_configured": true, 00:12:47.123 "data_offset": 2048, 00:12:47.123 "data_size": 63488 00:12:47.123 } 00:12:47.123 ] 00:12:47.123 }' 00:12:47.123 15:17:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:47.123 15:17:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:47.123 15:17:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:47.123 15:17:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:47.123 15:17:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:48.503 15:17:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:48.503 15:17:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:48.503 15:17:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:48.503 15:17:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:48.503 15:17:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:48.503 15:17:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:48.503 15:17:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:48.503 15:17:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:48.503 15:17:16 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:48.503 15:17:16 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:48.503 15:17:16 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:48.503 15:17:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:48.503 "name": "raid_bdev1", 00:12:48.503 "uuid": "2fa6bfe2-c1e7-4aed-bdfe-8d3d58cd3706", 00:12:48.503 "strip_size_kb": 0, 00:12:48.503 "state": "online", 00:12:48.503 "raid_level": "raid1", 00:12:48.503 "superblock": true, 00:12:48.503 "num_base_bdevs": 4, 00:12:48.503 "num_base_bdevs_discovered": 3, 00:12:48.503 "num_base_bdevs_operational": 3, 00:12:48.503 "process": { 00:12:48.503 "type": "rebuild", 00:12:48.503 "target": "spare", 00:12:48.503 "progress": { 00:12:48.503 "blocks": 51200, 00:12:48.503 "percent": 80 00:12:48.503 } 00:12:48.503 }, 00:12:48.503 "base_bdevs_list": [ 00:12:48.503 { 00:12:48.503 "name": "spare", 00:12:48.503 "uuid": "fd52ca22-1d55-53ff-b94e-f7e46e4b5f87", 00:12:48.503 "is_configured": true, 00:12:48.503 "data_offset": 2048, 00:12:48.503 "data_size": 63488 00:12:48.503 }, 00:12:48.503 { 00:12:48.503 "name": null, 00:12:48.503 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:48.503 "is_configured": false, 00:12:48.504 "data_offset": 0, 00:12:48.504 "data_size": 63488 00:12:48.504 }, 00:12:48.504 { 00:12:48.504 "name": "BaseBdev3", 00:12:48.504 "uuid": "3971710a-3a38-55c4-9cb6-8c93f5904fd5", 00:12:48.504 "is_configured": true, 00:12:48.504 "data_offset": 2048, 00:12:48.504 "data_size": 63488 00:12:48.504 }, 00:12:48.504 { 00:12:48.504 "name": "BaseBdev4", 00:12:48.504 "uuid": "71a325e4-1ec4-580c-ad64-3d2109daa198", 00:12:48.504 "is_configured": true, 00:12:48.504 "data_offset": 2048, 00:12:48.504 "data_size": 63488 00:12:48.504 } 00:12:48.504 ] 00:12:48.504 }' 00:12:48.504 15:17:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:48.504 15:17:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:48.504 15:17:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:48.504 15:17:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:48.504 15:17:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:48.763 [2024-11-27 15:17:16.786043] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:12:48.763 [2024-11-27 15:17:16.786130] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:12:48.763 [2024-11-27 15:17:16.786254] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:49.342 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:49.342 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:49.342 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:49.342 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:49.342 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:49.342 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:49.342 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:49.342 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:49.342 15:17:17 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:49.342 15:17:17 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:49.342 15:17:17 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:49.342 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:49.342 "name": "raid_bdev1", 00:12:49.342 "uuid": "2fa6bfe2-c1e7-4aed-bdfe-8d3d58cd3706", 00:12:49.342 "strip_size_kb": 0, 00:12:49.342 "state": "online", 00:12:49.342 "raid_level": "raid1", 00:12:49.342 "superblock": true, 00:12:49.342 "num_base_bdevs": 4, 00:12:49.342 "num_base_bdevs_discovered": 3, 00:12:49.342 "num_base_bdevs_operational": 3, 00:12:49.342 "base_bdevs_list": [ 00:12:49.342 { 00:12:49.342 "name": "spare", 00:12:49.342 "uuid": "fd52ca22-1d55-53ff-b94e-f7e46e4b5f87", 00:12:49.342 "is_configured": true, 00:12:49.342 "data_offset": 2048, 00:12:49.342 "data_size": 63488 00:12:49.342 }, 00:12:49.342 { 00:12:49.342 "name": null, 00:12:49.342 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:49.342 "is_configured": false, 00:12:49.342 "data_offset": 0, 00:12:49.342 "data_size": 63488 00:12:49.342 }, 00:12:49.342 { 00:12:49.342 "name": "BaseBdev3", 00:12:49.342 "uuid": "3971710a-3a38-55c4-9cb6-8c93f5904fd5", 00:12:49.342 "is_configured": true, 00:12:49.342 "data_offset": 2048, 00:12:49.342 "data_size": 63488 00:12:49.342 }, 00:12:49.342 { 00:12:49.342 "name": "BaseBdev4", 00:12:49.342 "uuid": "71a325e4-1ec4-580c-ad64-3d2109daa198", 00:12:49.342 "is_configured": true, 00:12:49.342 "data_offset": 2048, 00:12:49.342 "data_size": 63488 00:12:49.342 } 00:12:49.342 ] 00:12:49.342 }' 00:12:49.342 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:49.607 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:12:49.607 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:49.607 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:12:49.607 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:12:49.607 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:49.607 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:49.607 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:49.607 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:49.607 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:49.607 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:49.607 15:17:17 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:49.607 15:17:17 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:49.607 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:49.607 15:17:17 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:49.607 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:49.607 "name": "raid_bdev1", 00:12:49.607 "uuid": "2fa6bfe2-c1e7-4aed-bdfe-8d3d58cd3706", 00:12:49.607 "strip_size_kb": 0, 00:12:49.607 "state": "online", 00:12:49.607 "raid_level": "raid1", 00:12:49.607 "superblock": true, 00:12:49.607 "num_base_bdevs": 4, 00:12:49.607 "num_base_bdevs_discovered": 3, 00:12:49.607 "num_base_bdevs_operational": 3, 00:12:49.608 "base_bdevs_list": [ 00:12:49.608 { 00:12:49.608 "name": "spare", 00:12:49.608 "uuid": "fd52ca22-1d55-53ff-b94e-f7e46e4b5f87", 00:12:49.608 "is_configured": true, 00:12:49.608 "data_offset": 2048, 00:12:49.608 "data_size": 63488 00:12:49.608 }, 00:12:49.608 { 00:12:49.608 "name": null, 00:12:49.608 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:49.608 "is_configured": false, 00:12:49.608 "data_offset": 0, 00:12:49.608 "data_size": 63488 00:12:49.608 }, 00:12:49.608 { 00:12:49.608 "name": "BaseBdev3", 00:12:49.608 "uuid": "3971710a-3a38-55c4-9cb6-8c93f5904fd5", 00:12:49.608 "is_configured": true, 00:12:49.608 "data_offset": 2048, 00:12:49.608 "data_size": 63488 00:12:49.608 }, 00:12:49.608 { 00:12:49.608 "name": "BaseBdev4", 00:12:49.608 "uuid": "71a325e4-1ec4-580c-ad64-3d2109daa198", 00:12:49.608 "is_configured": true, 00:12:49.608 "data_offset": 2048, 00:12:49.608 "data_size": 63488 00:12:49.608 } 00:12:49.608 ] 00:12:49.608 }' 00:12:49.608 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:49.608 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:49.608 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:49.608 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:49.608 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:49.608 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:49.608 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:49.608 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:49.608 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:49.608 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:49.608 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:49.608 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:49.608 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:49.608 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:49.608 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:49.608 15:17:17 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:49.608 15:17:17 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:49.608 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:49.608 15:17:17 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:49.608 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:49.608 "name": "raid_bdev1", 00:12:49.608 "uuid": "2fa6bfe2-c1e7-4aed-bdfe-8d3d58cd3706", 00:12:49.608 "strip_size_kb": 0, 00:12:49.608 "state": "online", 00:12:49.608 "raid_level": "raid1", 00:12:49.608 "superblock": true, 00:12:49.608 "num_base_bdevs": 4, 00:12:49.608 "num_base_bdevs_discovered": 3, 00:12:49.608 "num_base_bdevs_operational": 3, 00:12:49.608 "base_bdevs_list": [ 00:12:49.608 { 00:12:49.608 "name": "spare", 00:12:49.608 "uuid": "fd52ca22-1d55-53ff-b94e-f7e46e4b5f87", 00:12:49.608 "is_configured": true, 00:12:49.608 "data_offset": 2048, 00:12:49.608 "data_size": 63488 00:12:49.608 }, 00:12:49.608 { 00:12:49.608 "name": null, 00:12:49.608 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:49.608 "is_configured": false, 00:12:49.608 "data_offset": 0, 00:12:49.608 "data_size": 63488 00:12:49.608 }, 00:12:49.608 { 00:12:49.608 "name": "BaseBdev3", 00:12:49.608 "uuid": "3971710a-3a38-55c4-9cb6-8c93f5904fd5", 00:12:49.608 "is_configured": true, 00:12:49.608 "data_offset": 2048, 00:12:49.608 "data_size": 63488 00:12:49.608 }, 00:12:49.608 { 00:12:49.608 "name": "BaseBdev4", 00:12:49.608 "uuid": "71a325e4-1ec4-580c-ad64-3d2109daa198", 00:12:49.608 "is_configured": true, 00:12:49.608 "data_offset": 2048, 00:12:49.608 "data_size": 63488 00:12:49.608 } 00:12:49.608 ] 00:12:49.608 }' 00:12:49.608 15:17:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:49.608 15:17:17 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:50.176 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:50.176 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:50.176 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:50.176 [2024-11-27 15:17:18.084487] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:50.176 [2024-11-27 15:17:18.084565] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:50.176 [2024-11-27 15:17:18.084731] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:50.176 [2024-11-27 15:17:18.084846] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:50.176 [2024-11-27 15:17:18.084916] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:12:50.176 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:50.176 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:50.176 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:50.176 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:50.176 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:12:50.176 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:50.176 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:12:50.176 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:12:50.177 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:12:50.177 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:12:50.177 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:50.177 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:12:50.177 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:50.177 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:50.177 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:50.177 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:12:50.177 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:50.177 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:50.177 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:12:50.436 /dev/nbd0 00:12:50.436 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:50.436 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:50.436 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:12:50.436 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:12:50.436 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:50.436 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:50.436 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:12:50.436 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:12:50.436 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:50.436 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:50.436 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:50.436 1+0 records in 00:12:50.436 1+0 records out 00:12:50.436 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000366325 s, 11.2 MB/s 00:12:50.436 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:50.436 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:12:50.436 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:50.436 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:50.436 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:12:50.436 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:50.436 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:50.436 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:12:50.697 /dev/nbd1 00:12:50.697 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:50.697 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:50.697 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:12:50.697 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:12:50.697 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:50.697 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:50.697 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:12:50.697 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:12:50.697 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:50.697 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:50.697 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:50.697 1+0 records in 00:12:50.697 1+0 records out 00:12:50.697 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000389583 s, 10.5 MB/s 00:12:50.697 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:50.697 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:12:50.697 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:50.697 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:50.697 15:17:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:12:50.697 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:50.697 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:50.697 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:12:50.697 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:12:50.697 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:50.697 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:50.697 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:50.697 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:12:50.697 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:50.697 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:50.956 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:50.956 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:50.956 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:50.956 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:50.956 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:50.956 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:50.956 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:12:50.956 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:12:50.956 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:50.956 15:17:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:51.227 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:51.227 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:51.227 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:51.227 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:51.227 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:51.227 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:51.227 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:12:51.227 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:12:51.227 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:12:51.227 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:12:51.227 15:17:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.227 15:17:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:51.227 15:17:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:51.227 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:51.227 15:17:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.227 15:17:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:51.227 [2024-11-27 15:17:19.120090] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:51.227 [2024-11-27 15:17:19.120211] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:51.227 [2024-11-27 15:17:19.120257] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:12:51.227 [2024-11-27 15:17:19.120308] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:51.227 [2024-11-27 15:17:19.122600] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:51.227 [2024-11-27 15:17:19.122680] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:51.227 [2024-11-27 15:17:19.122817] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:12:51.227 [2024-11-27 15:17:19.122925] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:51.227 [2024-11-27 15:17:19.123087] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:51.227 [2024-11-27 15:17:19.123234] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:51.227 spare 00:12:51.227 15:17:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:51.227 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:12:51.228 15:17:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.228 15:17:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:51.228 [2024-11-27 15:17:19.223175] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006600 00:12:51.228 [2024-11-27 15:17:19.223270] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:51.228 [2024-11-27 15:17:19.223660] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1ae0 00:12:51.228 [2024-11-27 15:17:19.223925] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006600 00:12:51.228 [2024-11-27 15:17:19.223973] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006600 00:12:51.228 [2024-11-27 15:17:19.224214] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:51.228 15:17:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:51.228 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:51.228 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:51.228 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:51.228 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:51.228 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:51.228 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:51.228 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:51.228 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:51.228 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:51.228 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:51.228 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:51.228 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:51.228 15:17:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.228 15:17:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:51.228 15:17:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:51.228 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:51.228 "name": "raid_bdev1", 00:12:51.228 "uuid": "2fa6bfe2-c1e7-4aed-bdfe-8d3d58cd3706", 00:12:51.228 "strip_size_kb": 0, 00:12:51.228 "state": "online", 00:12:51.228 "raid_level": "raid1", 00:12:51.228 "superblock": true, 00:12:51.228 "num_base_bdevs": 4, 00:12:51.228 "num_base_bdevs_discovered": 3, 00:12:51.228 "num_base_bdevs_operational": 3, 00:12:51.228 "base_bdevs_list": [ 00:12:51.228 { 00:12:51.228 "name": "spare", 00:12:51.228 "uuid": "fd52ca22-1d55-53ff-b94e-f7e46e4b5f87", 00:12:51.228 "is_configured": true, 00:12:51.228 "data_offset": 2048, 00:12:51.228 "data_size": 63488 00:12:51.228 }, 00:12:51.228 { 00:12:51.228 "name": null, 00:12:51.228 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:51.228 "is_configured": false, 00:12:51.228 "data_offset": 2048, 00:12:51.228 "data_size": 63488 00:12:51.228 }, 00:12:51.228 { 00:12:51.228 "name": "BaseBdev3", 00:12:51.228 "uuid": "3971710a-3a38-55c4-9cb6-8c93f5904fd5", 00:12:51.228 "is_configured": true, 00:12:51.228 "data_offset": 2048, 00:12:51.228 "data_size": 63488 00:12:51.228 }, 00:12:51.228 { 00:12:51.228 "name": "BaseBdev4", 00:12:51.228 "uuid": "71a325e4-1ec4-580c-ad64-3d2109daa198", 00:12:51.228 "is_configured": true, 00:12:51.228 "data_offset": 2048, 00:12:51.228 "data_size": 63488 00:12:51.228 } 00:12:51.228 ] 00:12:51.228 }' 00:12:51.228 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:51.228 15:17:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:51.796 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:51.796 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:51.796 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:51.796 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:51.796 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:51.796 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:51.796 15:17:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.796 15:17:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:51.796 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:51.796 15:17:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:51.796 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:51.796 "name": "raid_bdev1", 00:12:51.796 "uuid": "2fa6bfe2-c1e7-4aed-bdfe-8d3d58cd3706", 00:12:51.796 "strip_size_kb": 0, 00:12:51.796 "state": "online", 00:12:51.796 "raid_level": "raid1", 00:12:51.796 "superblock": true, 00:12:51.796 "num_base_bdevs": 4, 00:12:51.796 "num_base_bdevs_discovered": 3, 00:12:51.796 "num_base_bdevs_operational": 3, 00:12:51.796 "base_bdevs_list": [ 00:12:51.796 { 00:12:51.796 "name": "spare", 00:12:51.796 "uuid": "fd52ca22-1d55-53ff-b94e-f7e46e4b5f87", 00:12:51.796 "is_configured": true, 00:12:51.796 "data_offset": 2048, 00:12:51.796 "data_size": 63488 00:12:51.796 }, 00:12:51.796 { 00:12:51.796 "name": null, 00:12:51.796 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:51.796 "is_configured": false, 00:12:51.796 "data_offset": 2048, 00:12:51.796 "data_size": 63488 00:12:51.796 }, 00:12:51.796 { 00:12:51.796 "name": "BaseBdev3", 00:12:51.796 "uuid": "3971710a-3a38-55c4-9cb6-8c93f5904fd5", 00:12:51.796 "is_configured": true, 00:12:51.796 "data_offset": 2048, 00:12:51.796 "data_size": 63488 00:12:51.796 }, 00:12:51.796 { 00:12:51.796 "name": "BaseBdev4", 00:12:51.796 "uuid": "71a325e4-1ec4-580c-ad64-3d2109daa198", 00:12:51.796 "is_configured": true, 00:12:51.796 "data_offset": 2048, 00:12:51.796 "data_size": 63488 00:12:51.796 } 00:12:51.796 ] 00:12:51.796 }' 00:12:51.796 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:51.796 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:51.796 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:51.796 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:51.796 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:51.796 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:12:51.796 15:17:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.796 15:17:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:51.796 15:17:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:51.797 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:12:51.797 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:51.797 15:17:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.797 15:17:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:52.055 [2024-11-27 15:17:19.903290] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:52.055 15:17:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:52.055 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:52.055 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:52.055 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:52.055 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:52.055 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:52.055 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:52.055 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:52.055 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:52.055 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:52.055 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:52.055 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:52.055 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:52.055 15:17:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:52.055 15:17:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:52.055 15:17:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:52.055 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:52.055 "name": "raid_bdev1", 00:12:52.055 "uuid": "2fa6bfe2-c1e7-4aed-bdfe-8d3d58cd3706", 00:12:52.055 "strip_size_kb": 0, 00:12:52.055 "state": "online", 00:12:52.055 "raid_level": "raid1", 00:12:52.055 "superblock": true, 00:12:52.055 "num_base_bdevs": 4, 00:12:52.055 "num_base_bdevs_discovered": 2, 00:12:52.055 "num_base_bdevs_operational": 2, 00:12:52.055 "base_bdevs_list": [ 00:12:52.055 { 00:12:52.055 "name": null, 00:12:52.055 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:52.055 "is_configured": false, 00:12:52.055 "data_offset": 0, 00:12:52.055 "data_size": 63488 00:12:52.055 }, 00:12:52.055 { 00:12:52.055 "name": null, 00:12:52.055 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:52.055 "is_configured": false, 00:12:52.055 "data_offset": 2048, 00:12:52.055 "data_size": 63488 00:12:52.055 }, 00:12:52.055 { 00:12:52.055 "name": "BaseBdev3", 00:12:52.055 "uuid": "3971710a-3a38-55c4-9cb6-8c93f5904fd5", 00:12:52.055 "is_configured": true, 00:12:52.055 "data_offset": 2048, 00:12:52.055 "data_size": 63488 00:12:52.055 }, 00:12:52.055 { 00:12:52.055 "name": "BaseBdev4", 00:12:52.055 "uuid": "71a325e4-1ec4-580c-ad64-3d2109daa198", 00:12:52.055 "is_configured": true, 00:12:52.055 "data_offset": 2048, 00:12:52.055 "data_size": 63488 00:12:52.055 } 00:12:52.055 ] 00:12:52.055 }' 00:12:52.055 15:17:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:52.055 15:17:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:52.315 15:17:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:52.315 15:17:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:52.315 15:17:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:52.315 [2024-11-27 15:17:20.358653] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:52.315 [2024-11-27 15:17:20.359053] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:12:52.315 [2024-11-27 15:17:20.359088] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:12:52.315 [2024-11-27 15:17:20.359146] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:52.315 [2024-11-27 15:17:20.366413] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1bb0 00:12:52.315 15:17:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:52.315 15:17:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:12:52.315 [2024-11-27 15:17:20.368717] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:53.698 "name": "raid_bdev1", 00:12:53.698 "uuid": "2fa6bfe2-c1e7-4aed-bdfe-8d3d58cd3706", 00:12:53.698 "strip_size_kb": 0, 00:12:53.698 "state": "online", 00:12:53.698 "raid_level": "raid1", 00:12:53.698 "superblock": true, 00:12:53.698 "num_base_bdevs": 4, 00:12:53.698 "num_base_bdevs_discovered": 3, 00:12:53.698 "num_base_bdevs_operational": 3, 00:12:53.698 "process": { 00:12:53.698 "type": "rebuild", 00:12:53.698 "target": "spare", 00:12:53.698 "progress": { 00:12:53.698 "blocks": 20480, 00:12:53.698 "percent": 32 00:12:53.698 } 00:12:53.698 }, 00:12:53.698 "base_bdevs_list": [ 00:12:53.698 { 00:12:53.698 "name": "spare", 00:12:53.698 "uuid": "fd52ca22-1d55-53ff-b94e-f7e46e4b5f87", 00:12:53.698 "is_configured": true, 00:12:53.698 "data_offset": 2048, 00:12:53.698 "data_size": 63488 00:12:53.698 }, 00:12:53.698 { 00:12:53.698 "name": null, 00:12:53.698 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:53.698 "is_configured": false, 00:12:53.698 "data_offset": 2048, 00:12:53.698 "data_size": 63488 00:12:53.698 }, 00:12:53.698 { 00:12:53.698 "name": "BaseBdev3", 00:12:53.698 "uuid": "3971710a-3a38-55c4-9cb6-8c93f5904fd5", 00:12:53.698 "is_configured": true, 00:12:53.698 "data_offset": 2048, 00:12:53.698 "data_size": 63488 00:12:53.698 }, 00:12:53.698 { 00:12:53.698 "name": "BaseBdev4", 00:12:53.698 "uuid": "71a325e4-1ec4-580c-ad64-3d2109daa198", 00:12:53.698 "is_configured": true, 00:12:53.698 "data_offset": 2048, 00:12:53.698 "data_size": 63488 00:12:53.698 } 00:12:53.698 ] 00:12:53.698 }' 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:53.698 [2024-11-27 15:17:21.532675] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:53.698 [2024-11-27 15:17:21.578097] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:53.698 [2024-11-27 15:17:21.578253] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:53.698 [2024-11-27 15:17:21.578296] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:53.698 [2024-11-27 15:17:21.578324] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:53.698 "name": "raid_bdev1", 00:12:53.698 "uuid": "2fa6bfe2-c1e7-4aed-bdfe-8d3d58cd3706", 00:12:53.698 "strip_size_kb": 0, 00:12:53.698 "state": "online", 00:12:53.698 "raid_level": "raid1", 00:12:53.698 "superblock": true, 00:12:53.698 "num_base_bdevs": 4, 00:12:53.698 "num_base_bdevs_discovered": 2, 00:12:53.698 "num_base_bdevs_operational": 2, 00:12:53.698 "base_bdevs_list": [ 00:12:53.698 { 00:12:53.698 "name": null, 00:12:53.698 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:53.698 "is_configured": false, 00:12:53.698 "data_offset": 0, 00:12:53.698 "data_size": 63488 00:12:53.698 }, 00:12:53.698 { 00:12:53.698 "name": null, 00:12:53.698 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:53.698 "is_configured": false, 00:12:53.698 "data_offset": 2048, 00:12:53.698 "data_size": 63488 00:12:53.698 }, 00:12:53.698 { 00:12:53.698 "name": "BaseBdev3", 00:12:53.698 "uuid": "3971710a-3a38-55c4-9cb6-8c93f5904fd5", 00:12:53.698 "is_configured": true, 00:12:53.698 "data_offset": 2048, 00:12:53.698 "data_size": 63488 00:12:53.698 }, 00:12:53.698 { 00:12:53.698 "name": "BaseBdev4", 00:12:53.698 "uuid": "71a325e4-1ec4-580c-ad64-3d2109daa198", 00:12:53.698 "is_configured": true, 00:12:53.698 "data_offset": 2048, 00:12:53.698 "data_size": 63488 00:12:53.698 } 00:12:53.698 ] 00:12:53.698 }' 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:53.698 15:17:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:53.958 15:17:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:53.958 15:17:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.958 15:17:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:53.958 [2024-11-27 15:17:22.045709] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:53.958 [2024-11-27 15:17:22.045886] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:53.958 [2024-11-27 15:17:22.045960] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c380 00:12:53.958 [2024-11-27 15:17:22.046016] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:53.958 [2024-11-27 15:17:22.046631] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:53.958 [2024-11-27 15:17:22.046733] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:53.958 [2024-11-27 15:17:22.046894] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:12:53.958 [2024-11-27 15:17:22.046968] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:12:53.958 [2024-11-27 15:17:22.047020] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:12:53.958 [2024-11-27 15:17:22.047083] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:53.958 spare 00:12:53.958 [2024-11-27 15:17:22.054290] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1c80 00:12:53.958 15:17:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.958 15:17:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:12:53.958 [2024-11-27 15:17:22.056494] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:55.341 "name": "raid_bdev1", 00:12:55.341 "uuid": "2fa6bfe2-c1e7-4aed-bdfe-8d3d58cd3706", 00:12:55.341 "strip_size_kb": 0, 00:12:55.341 "state": "online", 00:12:55.341 "raid_level": "raid1", 00:12:55.341 "superblock": true, 00:12:55.341 "num_base_bdevs": 4, 00:12:55.341 "num_base_bdevs_discovered": 3, 00:12:55.341 "num_base_bdevs_operational": 3, 00:12:55.341 "process": { 00:12:55.341 "type": "rebuild", 00:12:55.341 "target": "spare", 00:12:55.341 "progress": { 00:12:55.341 "blocks": 20480, 00:12:55.341 "percent": 32 00:12:55.341 } 00:12:55.341 }, 00:12:55.341 "base_bdevs_list": [ 00:12:55.341 { 00:12:55.341 "name": "spare", 00:12:55.341 "uuid": "fd52ca22-1d55-53ff-b94e-f7e46e4b5f87", 00:12:55.341 "is_configured": true, 00:12:55.341 "data_offset": 2048, 00:12:55.341 "data_size": 63488 00:12:55.341 }, 00:12:55.341 { 00:12:55.341 "name": null, 00:12:55.341 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:55.341 "is_configured": false, 00:12:55.341 "data_offset": 2048, 00:12:55.341 "data_size": 63488 00:12:55.341 }, 00:12:55.341 { 00:12:55.341 "name": "BaseBdev3", 00:12:55.341 "uuid": "3971710a-3a38-55c4-9cb6-8c93f5904fd5", 00:12:55.341 "is_configured": true, 00:12:55.341 "data_offset": 2048, 00:12:55.341 "data_size": 63488 00:12:55.341 }, 00:12:55.341 { 00:12:55.341 "name": "BaseBdev4", 00:12:55.341 "uuid": "71a325e4-1ec4-580c-ad64-3d2109daa198", 00:12:55.341 "is_configured": true, 00:12:55.341 "data_offset": 2048, 00:12:55.341 "data_size": 63488 00:12:55.341 } 00:12:55.341 ] 00:12:55.341 }' 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:55.341 [2024-11-27 15:17:23.196552] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:55.341 [2024-11-27 15:17:23.265181] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:55.341 [2024-11-27 15:17:23.265303] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:55.341 [2024-11-27 15:17:23.265349] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:55.341 [2024-11-27 15:17:23.265374] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:55.341 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:55.341 "name": "raid_bdev1", 00:12:55.341 "uuid": "2fa6bfe2-c1e7-4aed-bdfe-8d3d58cd3706", 00:12:55.341 "strip_size_kb": 0, 00:12:55.341 "state": "online", 00:12:55.341 "raid_level": "raid1", 00:12:55.341 "superblock": true, 00:12:55.341 "num_base_bdevs": 4, 00:12:55.341 "num_base_bdevs_discovered": 2, 00:12:55.341 "num_base_bdevs_operational": 2, 00:12:55.341 "base_bdevs_list": [ 00:12:55.341 { 00:12:55.341 "name": null, 00:12:55.341 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:55.341 "is_configured": false, 00:12:55.341 "data_offset": 0, 00:12:55.341 "data_size": 63488 00:12:55.341 }, 00:12:55.341 { 00:12:55.341 "name": null, 00:12:55.341 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:55.341 "is_configured": false, 00:12:55.341 "data_offset": 2048, 00:12:55.341 "data_size": 63488 00:12:55.341 }, 00:12:55.341 { 00:12:55.341 "name": "BaseBdev3", 00:12:55.341 "uuid": "3971710a-3a38-55c4-9cb6-8c93f5904fd5", 00:12:55.341 "is_configured": true, 00:12:55.341 "data_offset": 2048, 00:12:55.341 "data_size": 63488 00:12:55.341 }, 00:12:55.341 { 00:12:55.341 "name": "BaseBdev4", 00:12:55.341 "uuid": "71a325e4-1ec4-580c-ad64-3d2109daa198", 00:12:55.341 "is_configured": true, 00:12:55.341 "data_offset": 2048, 00:12:55.341 "data_size": 63488 00:12:55.341 } 00:12:55.341 ] 00:12:55.342 }' 00:12:55.342 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:55.342 15:17:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:55.911 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:55.911 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:55.911 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:55.911 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:55.911 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:55.911 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:55.911 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:55.911 15:17:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:55.911 15:17:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:55.911 15:17:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:55.911 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:55.911 "name": "raid_bdev1", 00:12:55.911 "uuid": "2fa6bfe2-c1e7-4aed-bdfe-8d3d58cd3706", 00:12:55.911 "strip_size_kb": 0, 00:12:55.911 "state": "online", 00:12:55.911 "raid_level": "raid1", 00:12:55.911 "superblock": true, 00:12:55.911 "num_base_bdevs": 4, 00:12:55.911 "num_base_bdevs_discovered": 2, 00:12:55.911 "num_base_bdevs_operational": 2, 00:12:55.911 "base_bdevs_list": [ 00:12:55.911 { 00:12:55.911 "name": null, 00:12:55.911 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:55.911 "is_configured": false, 00:12:55.911 "data_offset": 0, 00:12:55.911 "data_size": 63488 00:12:55.911 }, 00:12:55.911 { 00:12:55.911 "name": null, 00:12:55.911 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:55.911 "is_configured": false, 00:12:55.911 "data_offset": 2048, 00:12:55.911 "data_size": 63488 00:12:55.911 }, 00:12:55.911 { 00:12:55.911 "name": "BaseBdev3", 00:12:55.911 "uuid": "3971710a-3a38-55c4-9cb6-8c93f5904fd5", 00:12:55.911 "is_configured": true, 00:12:55.911 "data_offset": 2048, 00:12:55.911 "data_size": 63488 00:12:55.911 }, 00:12:55.911 { 00:12:55.911 "name": "BaseBdev4", 00:12:55.911 "uuid": "71a325e4-1ec4-580c-ad64-3d2109daa198", 00:12:55.911 "is_configured": true, 00:12:55.911 "data_offset": 2048, 00:12:55.911 "data_size": 63488 00:12:55.911 } 00:12:55.911 ] 00:12:55.911 }' 00:12:55.911 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:55.911 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:55.911 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:55.911 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:55.911 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:12:55.911 15:17:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:55.911 15:17:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:55.911 15:17:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:55.911 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:55.911 15:17:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:55.911 15:17:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:55.911 [2024-11-27 15:17:23.927494] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:55.911 [2024-11-27 15:17:23.927660] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:55.911 [2024-11-27 15:17:23.927695] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c980 00:12:55.911 [2024-11-27 15:17:23.927706] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:55.911 [2024-11-27 15:17:23.928300] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:55.911 [2024-11-27 15:17:23.928327] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:55.911 [2024-11-27 15:17:23.928428] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:12:55.911 [2024-11-27 15:17:23.928445] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:12:55.911 [2024-11-27 15:17:23.928458] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:12:55.911 [2024-11-27 15:17:23.928473] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:12:55.911 BaseBdev1 00:12:55.911 15:17:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:55.911 15:17:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:12:56.849 15:17:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:56.849 15:17:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:56.849 15:17:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:56.849 15:17:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:56.849 15:17:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:56.849 15:17:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:56.849 15:17:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:56.849 15:17:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:56.849 15:17:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:56.849 15:17:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:56.849 15:17:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:56.849 15:17:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:56.850 15:17:24 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:56.850 15:17:24 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:57.110 15:17:24 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:57.110 15:17:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:57.110 "name": "raid_bdev1", 00:12:57.110 "uuid": "2fa6bfe2-c1e7-4aed-bdfe-8d3d58cd3706", 00:12:57.110 "strip_size_kb": 0, 00:12:57.110 "state": "online", 00:12:57.110 "raid_level": "raid1", 00:12:57.110 "superblock": true, 00:12:57.110 "num_base_bdevs": 4, 00:12:57.110 "num_base_bdevs_discovered": 2, 00:12:57.110 "num_base_bdevs_operational": 2, 00:12:57.110 "base_bdevs_list": [ 00:12:57.110 { 00:12:57.110 "name": null, 00:12:57.110 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:57.110 "is_configured": false, 00:12:57.110 "data_offset": 0, 00:12:57.110 "data_size": 63488 00:12:57.110 }, 00:12:57.110 { 00:12:57.110 "name": null, 00:12:57.110 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:57.110 "is_configured": false, 00:12:57.110 "data_offset": 2048, 00:12:57.110 "data_size": 63488 00:12:57.110 }, 00:12:57.110 { 00:12:57.110 "name": "BaseBdev3", 00:12:57.110 "uuid": "3971710a-3a38-55c4-9cb6-8c93f5904fd5", 00:12:57.110 "is_configured": true, 00:12:57.110 "data_offset": 2048, 00:12:57.110 "data_size": 63488 00:12:57.110 }, 00:12:57.110 { 00:12:57.110 "name": "BaseBdev4", 00:12:57.110 "uuid": "71a325e4-1ec4-580c-ad64-3d2109daa198", 00:12:57.110 "is_configured": true, 00:12:57.110 "data_offset": 2048, 00:12:57.110 "data_size": 63488 00:12:57.110 } 00:12:57.110 ] 00:12:57.110 }' 00:12:57.110 15:17:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:57.110 15:17:24 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:57.369 15:17:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:57.369 15:17:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:57.369 15:17:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:57.369 15:17:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:57.369 15:17:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:57.369 15:17:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:57.369 15:17:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:57.369 15:17:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:57.369 15:17:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:57.369 15:17:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:57.629 15:17:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:57.629 "name": "raid_bdev1", 00:12:57.629 "uuid": "2fa6bfe2-c1e7-4aed-bdfe-8d3d58cd3706", 00:12:57.629 "strip_size_kb": 0, 00:12:57.629 "state": "online", 00:12:57.629 "raid_level": "raid1", 00:12:57.629 "superblock": true, 00:12:57.629 "num_base_bdevs": 4, 00:12:57.629 "num_base_bdevs_discovered": 2, 00:12:57.629 "num_base_bdevs_operational": 2, 00:12:57.629 "base_bdevs_list": [ 00:12:57.629 { 00:12:57.629 "name": null, 00:12:57.630 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:57.630 "is_configured": false, 00:12:57.630 "data_offset": 0, 00:12:57.630 "data_size": 63488 00:12:57.630 }, 00:12:57.630 { 00:12:57.630 "name": null, 00:12:57.630 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:57.630 "is_configured": false, 00:12:57.630 "data_offset": 2048, 00:12:57.630 "data_size": 63488 00:12:57.630 }, 00:12:57.630 { 00:12:57.630 "name": "BaseBdev3", 00:12:57.630 "uuid": "3971710a-3a38-55c4-9cb6-8c93f5904fd5", 00:12:57.630 "is_configured": true, 00:12:57.630 "data_offset": 2048, 00:12:57.630 "data_size": 63488 00:12:57.630 }, 00:12:57.630 { 00:12:57.630 "name": "BaseBdev4", 00:12:57.630 "uuid": "71a325e4-1ec4-580c-ad64-3d2109daa198", 00:12:57.630 "is_configured": true, 00:12:57.630 "data_offset": 2048, 00:12:57.630 "data_size": 63488 00:12:57.630 } 00:12:57.630 ] 00:12:57.630 }' 00:12:57.630 15:17:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:57.630 15:17:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:57.630 15:17:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:57.630 15:17:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:57.630 15:17:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:57.630 15:17:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@652 -- # local es=0 00:12:57.630 15:17:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:57.630 15:17:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:12:57.630 15:17:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:12:57.630 15:17:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:12:57.630 15:17:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:12:57.630 15:17:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:57.630 15:17:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:57.630 15:17:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:57.630 [2024-11-27 15:17:25.580764] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:57.630 [2024-11-27 15:17:25.581009] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:12:57.630 [2024-11-27 15:17:25.581085] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:12:57.630 request: 00:12:57.630 { 00:12:57.630 "base_bdev": "BaseBdev1", 00:12:57.630 "raid_bdev": "raid_bdev1", 00:12:57.630 "method": "bdev_raid_add_base_bdev", 00:12:57.630 "req_id": 1 00:12:57.630 } 00:12:57.630 Got JSON-RPC error response 00:12:57.630 response: 00:12:57.630 { 00:12:57.630 "code": -22, 00:12:57.630 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:12:57.630 } 00:12:57.630 15:17:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:12:57.630 15:17:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@655 -- # es=1 00:12:57.630 15:17:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:12:57.630 15:17:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:12:57.630 15:17:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:12:57.630 15:17:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:12:58.568 15:17:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:58.568 15:17:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:58.568 15:17:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:58.568 15:17:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:58.568 15:17:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:58.568 15:17:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:58.568 15:17:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:58.568 15:17:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:58.568 15:17:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:58.568 15:17:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:58.568 15:17:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:58.568 15:17:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:58.568 15:17:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:58.568 15:17:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:58.568 15:17:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:58.568 15:17:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:58.568 "name": "raid_bdev1", 00:12:58.568 "uuid": "2fa6bfe2-c1e7-4aed-bdfe-8d3d58cd3706", 00:12:58.568 "strip_size_kb": 0, 00:12:58.568 "state": "online", 00:12:58.568 "raid_level": "raid1", 00:12:58.568 "superblock": true, 00:12:58.568 "num_base_bdevs": 4, 00:12:58.568 "num_base_bdevs_discovered": 2, 00:12:58.568 "num_base_bdevs_operational": 2, 00:12:58.568 "base_bdevs_list": [ 00:12:58.568 { 00:12:58.568 "name": null, 00:12:58.568 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:58.568 "is_configured": false, 00:12:58.568 "data_offset": 0, 00:12:58.568 "data_size": 63488 00:12:58.568 }, 00:12:58.568 { 00:12:58.568 "name": null, 00:12:58.568 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:58.568 "is_configured": false, 00:12:58.568 "data_offset": 2048, 00:12:58.568 "data_size": 63488 00:12:58.568 }, 00:12:58.568 { 00:12:58.568 "name": "BaseBdev3", 00:12:58.568 "uuid": "3971710a-3a38-55c4-9cb6-8c93f5904fd5", 00:12:58.568 "is_configured": true, 00:12:58.568 "data_offset": 2048, 00:12:58.568 "data_size": 63488 00:12:58.568 }, 00:12:58.568 { 00:12:58.568 "name": "BaseBdev4", 00:12:58.568 "uuid": "71a325e4-1ec4-580c-ad64-3d2109daa198", 00:12:58.568 "is_configured": true, 00:12:58.568 "data_offset": 2048, 00:12:58.568 "data_size": 63488 00:12:58.568 } 00:12:58.568 ] 00:12:58.568 }' 00:12:58.568 15:17:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:58.568 15:17:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:59.137 15:17:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:59.137 15:17:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:59.137 15:17:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:59.137 15:17:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:59.137 15:17:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:59.137 15:17:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:59.137 15:17:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:59.137 15:17:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:59.137 15:17:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:59.137 15:17:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:59.137 15:17:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:59.137 "name": "raid_bdev1", 00:12:59.137 "uuid": "2fa6bfe2-c1e7-4aed-bdfe-8d3d58cd3706", 00:12:59.137 "strip_size_kb": 0, 00:12:59.137 "state": "online", 00:12:59.137 "raid_level": "raid1", 00:12:59.137 "superblock": true, 00:12:59.137 "num_base_bdevs": 4, 00:12:59.137 "num_base_bdevs_discovered": 2, 00:12:59.137 "num_base_bdevs_operational": 2, 00:12:59.137 "base_bdevs_list": [ 00:12:59.137 { 00:12:59.137 "name": null, 00:12:59.137 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:59.137 "is_configured": false, 00:12:59.137 "data_offset": 0, 00:12:59.137 "data_size": 63488 00:12:59.137 }, 00:12:59.137 { 00:12:59.137 "name": null, 00:12:59.137 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:59.137 "is_configured": false, 00:12:59.137 "data_offset": 2048, 00:12:59.137 "data_size": 63488 00:12:59.137 }, 00:12:59.137 { 00:12:59.137 "name": "BaseBdev3", 00:12:59.137 "uuid": "3971710a-3a38-55c4-9cb6-8c93f5904fd5", 00:12:59.137 "is_configured": true, 00:12:59.137 "data_offset": 2048, 00:12:59.137 "data_size": 63488 00:12:59.137 }, 00:12:59.137 { 00:12:59.137 "name": "BaseBdev4", 00:12:59.137 "uuid": "71a325e4-1ec4-580c-ad64-3d2109daa198", 00:12:59.137 "is_configured": true, 00:12:59.137 "data_offset": 2048, 00:12:59.137 "data_size": 63488 00:12:59.137 } 00:12:59.137 ] 00:12:59.137 }' 00:12:59.137 15:17:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:59.137 15:17:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:59.137 15:17:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:59.137 15:17:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:59.137 15:17:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 88793 00:12:59.137 15:17:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@954 -- # '[' -z 88793 ']' 00:12:59.137 15:17:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@958 -- # kill -0 88793 00:12:59.137 15:17:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@959 -- # uname 00:12:59.137 15:17:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:59.137 15:17:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 88793 00:12:59.137 15:17:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:59.137 15:17:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:59.137 killing process with pid 88793 00:12:59.137 Received shutdown signal, test time was about 60.000000 seconds 00:12:59.137 00:12:59.137 Latency(us) 00:12:59.137 [2024-11-27T15:17:27.244Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:59.137 [2024-11-27T15:17:27.244Z] =================================================================================================================== 00:12:59.137 [2024-11-27T15:17:27.244Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:12:59.137 15:17:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 88793' 00:12:59.137 15:17:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@973 -- # kill 88793 00:12:59.137 [2024-11-27 15:17:27.225882] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:59.137 15:17:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@978 -- # wait 88793 00:12:59.137 [2024-11-27 15:17:27.226047] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:59.137 [2024-11-27 15:17:27.226126] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:59.137 [2024-11-27 15:17:27.226139] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state offline 00:12:59.396 [2024-11-27 15:17:27.279002] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:59.396 15:17:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:12:59.396 00:12:59.396 real 0m23.740s 00:12:59.396 user 0m28.805s 00:12:59.396 sys 0m3.835s 00:12:59.396 15:17:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:59.396 15:17:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:59.396 ************************************ 00:12:59.396 END TEST raid_rebuild_test_sb 00:12:59.396 ************************************ 00:12:59.655 15:17:27 bdev_raid -- bdev/bdev_raid.sh@980 -- # run_test raid_rebuild_test_io raid_rebuild_test raid1 4 false true true 00:12:59.655 15:17:27 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:12:59.655 15:17:27 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:59.655 15:17:27 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:59.655 ************************************ 00:12:59.655 START TEST raid_rebuild_test_io 00:12:59.655 ************************************ 00:12:59.655 15:17:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 4 false true true 00:12:59.655 15:17:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:12:59.655 15:17:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:12:59.655 15:17:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:12:59.655 15:17:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:12:59.655 15:17:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:59.655 15:17:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:59.655 15:17:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:59.655 15:17:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:59.655 15:17:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:59.655 15:17:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:59.655 15:17:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:59.655 15:17:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:59.655 15:17:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:59.655 15:17:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:12:59.655 15:17:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:59.655 15:17:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:59.655 15:17:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:12:59.655 15:17:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:59.655 15:17:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:59.655 15:17:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:12:59.655 15:17:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:59.655 15:17:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:59.655 15:17:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:59.655 15:17:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:59.655 15:17:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:59.655 15:17:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:59.655 15:17:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:12:59.656 15:17:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:12:59.656 15:17:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:12:59.656 15:17:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@597 -- # raid_pid=89530 00:12:59.656 15:17:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 89530 00:12:59.656 15:17:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:59.656 15:17:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@835 -- # '[' -z 89530 ']' 00:12:59.656 15:17:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:59.656 15:17:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:59.656 15:17:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:59.656 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:59.656 15:17:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:59.656 15:17:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:59.656 [2024-11-27 15:17:27.652549] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:12:59.656 [2024-11-27 15:17:27.652782] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid89530 ] 00:12:59.656 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:59.656 Zero copy mechanism will not be used. 00:12:59.914 [2024-11-27 15:17:27.824382] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:59.914 [2024-11-27 15:17:27.850507] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:59.914 [2024-11-27 15:17:27.894732] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:59.914 [2024-11-27 15:17:27.894844] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:00.482 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:13:00.482 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # return 0 00:13:00.482 15:17:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:00.482 15:17:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:13:00.482 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:00.482 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:00.482 BaseBdev1_malloc 00:13:00.482 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:00.482 15:17:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:00.482 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:00.482 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:00.482 [2024-11-27 15:17:28.555638] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:00.482 [2024-11-27 15:17:28.555736] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:00.482 [2024-11-27 15:17:28.555780] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:13:00.482 [2024-11-27 15:17:28.555810] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:00.482 [2024-11-27 15:17:28.557874] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:00.482 [2024-11-27 15:17:28.557958] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:00.482 BaseBdev1 00:13:00.482 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:00.482 15:17:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:00.482 15:17:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:13:00.482 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:00.482 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:00.482 BaseBdev2_malloc 00:13:00.482 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:00.482 15:17:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:13:00.482 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:00.482 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:00.482 [2024-11-27 15:17:28.584450] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:13:00.482 [2024-11-27 15:17:28.584549] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:00.482 [2024-11-27 15:17:28.584604] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:13:00.482 [2024-11-27 15:17:28.584638] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:00.482 [2024-11-27 15:17:28.586744] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:00.482 [2024-11-27 15:17:28.586815] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:13:00.742 BaseBdev2 00:13:00.742 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:00.742 15:17:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:00.742 15:17:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:13:00.742 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:00.742 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:00.742 BaseBdev3_malloc 00:13:00.742 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:00.742 15:17:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:13:00.742 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:00.742 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:00.742 [2024-11-27 15:17:28.613252] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:13:00.742 [2024-11-27 15:17:28.613355] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:00.742 [2024-11-27 15:17:28.613394] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:13:00.742 [2024-11-27 15:17:28.613422] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:00.742 [2024-11-27 15:17:28.615406] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:00.742 [2024-11-27 15:17:28.615485] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:13:00.742 BaseBdev3 00:13:00.742 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:00.742 15:17:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:00.742 15:17:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:13:00.742 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:00.742 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:00.742 BaseBdev4_malloc 00:13:00.742 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:00.742 15:17:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:13:00.742 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:00.742 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:00.742 [2024-11-27 15:17:28.650073] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:13:00.742 [2024-11-27 15:17:28.650173] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:00.743 [2024-11-27 15:17:28.650219] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:13:00.743 [2024-11-27 15:17:28.650249] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:00.743 [2024-11-27 15:17:28.652328] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:00.743 [2024-11-27 15:17:28.652400] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:13:00.743 BaseBdev4 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:00.743 spare_malloc 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:00.743 spare_delay 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:00.743 [2024-11-27 15:17:28.690874] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:00.743 [2024-11-27 15:17:28.690974] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:00.743 [2024-11-27 15:17:28.691012] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:13:00.743 [2024-11-27 15:17:28.691024] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:00.743 [2024-11-27 15:17:28.693166] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:00.743 [2024-11-27 15:17:28.693204] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:00.743 spare 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:00.743 [2024-11-27 15:17:28.702942] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:00.743 [2024-11-27 15:17:28.704982] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:00.743 [2024-11-27 15:17:28.705093] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:00.743 [2024-11-27 15:17:28.705157] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:13:00.743 [2024-11-27 15:17:28.705269] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:13:00.743 [2024-11-27 15:17:28.705309] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:13:00.743 [2024-11-27 15:17:28.705603] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:13:00.743 [2024-11-27 15:17:28.705781] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:13:00.743 [2024-11-27 15:17:28.705830] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:13:00.743 [2024-11-27 15:17:28.706023] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:00.743 "name": "raid_bdev1", 00:13:00.743 "uuid": "bfafadfd-f3ca-409e-b748-7b5977cf95fe", 00:13:00.743 "strip_size_kb": 0, 00:13:00.743 "state": "online", 00:13:00.743 "raid_level": "raid1", 00:13:00.743 "superblock": false, 00:13:00.743 "num_base_bdevs": 4, 00:13:00.743 "num_base_bdevs_discovered": 4, 00:13:00.743 "num_base_bdevs_operational": 4, 00:13:00.743 "base_bdevs_list": [ 00:13:00.743 { 00:13:00.743 "name": "BaseBdev1", 00:13:00.743 "uuid": "75768f08-97c1-57c9-b915-23090ed0072c", 00:13:00.743 "is_configured": true, 00:13:00.743 "data_offset": 0, 00:13:00.743 "data_size": 65536 00:13:00.743 }, 00:13:00.743 { 00:13:00.743 "name": "BaseBdev2", 00:13:00.743 "uuid": "878e58ea-4dd1-5328-89d0-3824e6ebaf61", 00:13:00.743 "is_configured": true, 00:13:00.743 "data_offset": 0, 00:13:00.743 "data_size": 65536 00:13:00.743 }, 00:13:00.743 { 00:13:00.743 "name": "BaseBdev3", 00:13:00.743 "uuid": "09878590-9963-5efd-800a-1862515421c6", 00:13:00.743 "is_configured": true, 00:13:00.743 "data_offset": 0, 00:13:00.743 "data_size": 65536 00:13:00.743 }, 00:13:00.743 { 00:13:00.743 "name": "BaseBdev4", 00:13:00.743 "uuid": "6bfd443d-22ee-5db5-8f9b-d054694b31be", 00:13:00.743 "is_configured": true, 00:13:00.743 "data_offset": 0, 00:13:00.743 "data_size": 65536 00:13:00.743 } 00:13:00.743 ] 00:13:00.743 }' 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:00.743 15:17:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:01.311 [2024-11-27 15:17:29.142474] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:01.311 [2024-11-27 15:17:29.214043] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:01.311 "name": "raid_bdev1", 00:13:01.311 "uuid": "bfafadfd-f3ca-409e-b748-7b5977cf95fe", 00:13:01.311 "strip_size_kb": 0, 00:13:01.311 "state": "online", 00:13:01.311 "raid_level": "raid1", 00:13:01.311 "superblock": false, 00:13:01.311 "num_base_bdevs": 4, 00:13:01.311 "num_base_bdevs_discovered": 3, 00:13:01.311 "num_base_bdevs_operational": 3, 00:13:01.311 "base_bdevs_list": [ 00:13:01.311 { 00:13:01.311 "name": null, 00:13:01.311 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:01.311 "is_configured": false, 00:13:01.311 "data_offset": 0, 00:13:01.311 "data_size": 65536 00:13:01.311 }, 00:13:01.311 { 00:13:01.311 "name": "BaseBdev2", 00:13:01.311 "uuid": "878e58ea-4dd1-5328-89d0-3824e6ebaf61", 00:13:01.311 "is_configured": true, 00:13:01.311 "data_offset": 0, 00:13:01.311 "data_size": 65536 00:13:01.311 }, 00:13:01.311 { 00:13:01.311 "name": "BaseBdev3", 00:13:01.311 "uuid": "09878590-9963-5efd-800a-1862515421c6", 00:13:01.311 "is_configured": true, 00:13:01.311 "data_offset": 0, 00:13:01.311 "data_size": 65536 00:13:01.311 }, 00:13:01.311 { 00:13:01.311 "name": "BaseBdev4", 00:13:01.311 "uuid": "6bfd443d-22ee-5db5-8f9b-d054694b31be", 00:13:01.311 "is_configured": true, 00:13:01.311 "data_offset": 0, 00:13:01.311 "data_size": 65536 00:13:01.311 } 00:13:01.311 ] 00:13:01.311 }' 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:01.311 15:17:29 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:01.311 [2024-11-27 15:17:29.284055] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:13:01.311 I/O size of 3145728 is greater than zero copy threshold (65536). 00:13:01.311 Zero copy mechanism will not be used. 00:13:01.311 Running I/O for 60 seconds... 00:13:01.878 15:17:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:01.878 15:17:29 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:01.878 15:17:29 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:01.878 [2024-11-27 15:17:29.706057] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:01.878 15:17:29 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:01.878 15:17:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:13:01.878 [2024-11-27 15:17:29.756259] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:13:01.878 [2024-11-27 15:17:29.758296] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:01.878 [2024-11-27 15:17:29.874291] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:01.878 [2024-11-27 15:17:29.874826] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:02.137 [2024-11-27 15:17:30.007892] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:02.137 [2024-11-27 15:17:30.008318] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:02.394 196.00 IOPS, 588.00 MiB/s [2024-11-27T15:17:30.501Z] [2024-11-27 15:17:30.349456] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:13:02.394 [2024-11-27 15:17:30.350073] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:13:02.661 [2024-11-27 15:17:30.580963] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:02.661 [2024-11-27 15:17:30.581752] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:02.661 15:17:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:02.661 15:17:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:02.661 15:17:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:02.661 15:17:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:02.661 15:17:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:02.661 15:17:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:02.661 15:17:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:02.661 15:17:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:02.661 15:17:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:02.921 15:17:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:02.921 15:17:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:02.921 "name": "raid_bdev1", 00:13:02.921 "uuid": "bfafadfd-f3ca-409e-b748-7b5977cf95fe", 00:13:02.921 "strip_size_kb": 0, 00:13:02.921 "state": "online", 00:13:02.921 "raid_level": "raid1", 00:13:02.921 "superblock": false, 00:13:02.921 "num_base_bdevs": 4, 00:13:02.921 "num_base_bdevs_discovered": 4, 00:13:02.921 "num_base_bdevs_operational": 4, 00:13:02.921 "process": { 00:13:02.921 "type": "rebuild", 00:13:02.921 "target": "spare", 00:13:02.921 "progress": { 00:13:02.921 "blocks": 10240, 00:13:02.921 "percent": 15 00:13:02.921 } 00:13:02.921 }, 00:13:02.921 "base_bdevs_list": [ 00:13:02.921 { 00:13:02.921 "name": "spare", 00:13:02.921 "uuid": "3fe1edd0-f1a2-57df-9e90-ad085859747f", 00:13:02.921 "is_configured": true, 00:13:02.921 "data_offset": 0, 00:13:02.921 "data_size": 65536 00:13:02.921 }, 00:13:02.921 { 00:13:02.921 "name": "BaseBdev2", 00:13:02.921 "uuid": "878e58ea-4dd1-5328-89d0-3824e6ebaf61", 00:13:02.921 "is_configured": true, 00:13:02.921 "data_offset": 0, 00:13:02.921 "data_size": 65536 00:13:02.921 }, 00:13:02.921 { 00:13:02.921 "name": "BaseBdev3", 00:13:02.921 "uuid": "09878590-9963-5efd-800a-1862515421c6", 00:13:02.921 "is_configured": true, 00:13:02.921 "data_offset": 0, 00:13:02.921 "data_size": 65536 00:13:02.921 }, 00:13:02.921 { 00:13:02.921 "name": "BaseBdev4", 00:13:02.921 "uuid": "6bfd443d-22ee-5db5-8f9b-d054694b31be", 00:13:02.921 "is_configured": true, 00:13:02.921 "data_offset": 0, 00:13:02.921 "data_size": 65536 00:13:02.921 } 00:13:02.921 ] 00:13:02.921 }' 00:13:02.921 15:17:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:02.921 15:17:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:02.921 15:17:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:02.921 15:17:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:02.921 15:17:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:02.921 15:17:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:02.921 15:17:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:02.921 [2024-11-27 15:17:30.846578] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:02.921 [2024-11-27 15:17:30.902316] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:13:02.921 [2024-11-27 15:17:30.903642] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:13:02.921 [2024-11-27 15:17:31.005021] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:02.921 [2024-11-27 15:17:31.013439] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:02.921 [2024-11-27 15:17:31.013537] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:02.921 [2024-11-27 15:17:31.013563] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:03.181 [2024-11-27 15:17:31.026160] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000006080 00:13:03.181 15:17:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.181 15:17:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:03.181 15:17:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:03.181 15:17:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:03.181 15:17:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:03.181 15:17:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:03.181 15:17:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:03.181 15:17:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:03.181 15:17:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:03.181 15:17:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:03.181 15:17:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:03.181 15:17:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:03.181 15:17:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:03.181 15:17:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.181 15:17:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:03.181 15:17:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.181 15:17:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:03.181 "name": "raid_bdev1", 00:13:03.181 "uuid": "bfafadfd-f3ca-409e-b748-7b5977cf95fe", 00:13:03.181 "strip_size_kb": 0, 00:13:03.181 "state": "online", 00:13:03.181 "raid_level": "raid1", 00:13:03.181 "superblock": false, 00:13:03.181 "num_base_bdevs": 4, 00:13:03.181 "num_base_bdevs_discovered": 3, 00:13:03.181 "num_base_bdevs_operational": 3, 00:13:03.181 "base_bdevs_list": [ 00:13:03.181 { 00:13:03.181 "name": null, 00:13:03.181 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:03.181 "is_configured": false, 00:13:03.181 "data_offset": 0, 00:13:03.181 "data_size": 65536 00:13:03.181 }, 00:13:03.181 { 00:13:03.181 "name": "BaseBdev2", 00:13:03.181 "uuid": "878e58ea-4dd1-5328-89d0-3824e6ebaf61", 00:13:03.181 "is_configured": true, 00:13:03.181 "data_offset": 0, 00:13:03.181 "data_size": 65536 00:13:03.181 }, 00:13:03.181 { 00:13:03.181 "name": "BaseBdev3", 00:13:03.181 "uuid": "09878590-9963-5efd-800a-1862515421c6", 00:13:03.181 "is_configured": true, 00:13:03.181 "data_offset": 0, 00:13:03.181 "data_size": 65536 00:13:03.181 }, 00:13:03.181 { 00:13:03.181 "name": "BaseBdev4", 00:13:03.181 "uuid": "6bfd443d-22ee-5db5-8f9b-d054694b31be", 00:13:03.181 "is_configured": true, 00:13:03.181 "data_offset": 0, 00:13:03.181 "data_size": 65536 00:13:03.181 } 00:13:03.181 ] 00:13:03.181 }' 00:13:03.181 15:17:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:03.181 15:17:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:03.440 147.50 IOPS, 442.50 MiB/s [2024-11-27T15:17:31.547Z] 15:17:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:03.440 15:17:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:03.440 15:17:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:03.440 15:17:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:03.440 15:17:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:03.440 15:17:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:03.440 15:17:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:03.440 15:17:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.440 15:17:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:03.440 15:17:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.440 15:17:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:03.440 "name": "raid_bdev1", 00:13:03.440 "uuid": "bfafadfd-f3ca-409e-b748-7b5977cf95fe", 00:13:03.440 "strip_size_kb": 0, 00:13:03.440 "state": "online", 00:13:03.440 "raid_level": "raid1", 00:13:03.440 "superblock": false, 00:13:03.440 "num_base_bdevs": 4, 00:13:03.440 "num_base_bdevs_discovered": 3, 00:13:03.440 "num_base_bdevs_operational": 3, 00:13:03.440 "base_bdevs_list": [ 00:13:03.440 { 00:13:03.440 "name": null, 00:13:03.440 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:03.440 "is_configured": false, 00:13:03.440 "data_offset": 0, 00:13:03.440 "data_size": 65536 00:13:03.440 }, 00:13:03.440 { 00:13:03.440 "name": "BaseBdev2", 00:13:03.440 "uuid": "878e58ea-4dd1-5328-89d0-3824e6ebaf61", 00:13:03.440 "is_configured": true, 00:13:03.440 "data_offset": 0, 00:13:03.440 "data_size": 65536 00:13:03.440 }, 00:13:03.440 { 00:13:03.440 "name": "BaseBdev3", 00:13:03.440 "uuid": "09878590-9963-5efd-800a-1862515421c6", 00:13:03.440 "is_configured": true, 00:13:03.440 "data_offset": 0, 00:13:03.440 "data_size": 65536 00:13:03.440 }, 00:13:03.440 { 00:13:03.440 "name": "BaseBdev4", 00:13:03.440 "uuid": "6bfd443d-22ee-5db5-8f9b-d054694b31be", 00:13:03.440 "is_configured": true, 00:13:03.440 "data_offset": 0, 00:13:03.440 "data_size": 65536 00:13:03.440 } 00:13:03.440 ] 00:13:03.440 }' 00:13:03.440 15:17:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:03.699 15:17:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:03.699 15:17:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:03.699 15:17:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:03.699 15:17:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:03.699 15:17:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.699 15:17:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:03.699 [2024-11-27 15:17:31.622882] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:03.699 15:17:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.699 15:17:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:13:03.699 [2024-11-27 15:17:31.646788] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:13:03.699 [2024-11-27 15:17:31.648936] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:03.699 [2024-11-27 15:17:31.765490] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:03.957 [2024-11-27 15:17:31.890042] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:03.958 [2024-11-27 15:17:31.890303] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:04.476 167.00 IOPS, 501.00 MiB/s [2024-11-27T15:17:32.583Z] [2024-11-27 15:17:32.371922] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:04.476 [2024-11-27 15:17:32.372222] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:04.735 15:17:32 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:04.735 15:17:32 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:04.735 15:17:32 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:04.735 15:17:32 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:04.735 15:17:32 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:04.735 15:17:32 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:04.735 15:17:32 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:04.735 15:17:32 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:04.735 15:17:32 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:04.735 15:17:32 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:04.736 15:17:32 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:04.736 "name": "raid_bdev1", 00:13:04.736 "uuid": "bfafadfd-f3ca-409e-b748-7b5977cf95fe", 00:13:04.736 "strip_size_kb": 0, 00:13:04.736 "state": "online", 00:13:04.736 "raid_level": "raid1", 00:13:04.736 "superblock": false, 00:13:04.736 "num_base_bdevs": 4, 00:13:04.736 "num_base_bdevs_discovered": 4, 00:13:04.736 "num_base_bdevs_operational": 4, 00:13:04.736 "process": { 00:13:04.736 "type": "rebuild", 00:13:04.736 "target": "spare", 00:13:04.736 "progress": { 00:13:04.736 "blocks": 14336, 00:13:04.736 "percent": 21 00:13:04.736 } 00:13:04.736 }, 00:13:04.736 "base_bdevs_list": [ 00:13:04.736 { 00:13:04.736 "name": "spare", 00:13:04.736 "uuid": "3fe1edd0-f1a2-57df-9e90-ad085859747f", 00:13:04.736 "is_configured": true, 00:13:04.736 "data_offset": 0, 00:13:04.736 "data_size": 65536 00:13:04.736 }, 00:13:04.736 { 00:13:04.736 "name": "BaseBdev2", 00:13:04.736 "uuid": "878e58ea-4dd1-5328-89d0-3824e6ebaf61", 00:13:04.736 "is_configured": true, 00:13:04.736 "data_offset": 0, 00:13:04.736 "data_size": 65536 00:13:04.736 }, 00:13:04.736 { 00:13:04.736 "name": "BaseBdev3", 00:13:04.736 "uuid": "09878590-9963-5efd-800a-1862515421c6", 00:13:04.736 "is_configured": true, 00:13:04.736 "data_offset": 0, 00:13:04.736 "data_size": 65536 00:13:04.736 }, 00:13:04.736 { 00:13:04.736 "name": "BaseBdev4", 00:13:04.736 "uuid": "6bfd443d-22ee-5db5-8f9b-d054694b31be", 00:13:04.736 "is_configured": true, 00:13:04.736 "data_offset": 0, 00:13:04.736 "data_size": 65536 00:13:04.736 } 00:13:04.736 ] 00:13:04.736 }' 00:13:04.736 15:17:32 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:04.736 15:17:32 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:04.736 [2024-11-27 15:17:32.737565] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:13:04.736 15:17:32 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:04.736 15:17:32 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:04.736 15:17:32 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:13:04.736 15:17:32 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:13:04.736 15:17:32 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:13:04.736 15:17:32 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:13:04.736 15:17:32 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:13:04.736 15:17:32 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:04.736 15:17:32 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:04.736 [2024-11-27 15:17:32.794177] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:04.995 [2024-11-27 15:17:32.996872] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000006080 00:13:04.995 [2024-11-27 15:17:32.997012] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000006220 00:13:04.995 15:17:33 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:04.995 15:17:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:13:04.995 15:17:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:13:04.995 15:17:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:04.995 15:17:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:04.995 15:17:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:04.995 15:17:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:04.995 15:17:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:04.995 15:17:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:04.995 15:17:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:04.995 15:17:33 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:04.995 15:17:33 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:04.995 15:17:33 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:04.995 15:17:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:04.995 "name": "raid_bdev1", 00:13:04.995 "uuid": "bfafadfd-f3ca-409e-b748-7b5977cf95fe", 00:13:04.996 "strip_size_kb": 0, 00:13:04.996 "state": "online", 00:13:04.996 "raid_level": "raid1", 00:13:04.996 "superblock": false, 00:13:04.996 "num_base_bdevs": 4, 00:13:04.996 "num_base_bdevs_discovered": 3, 00:13:04.996 "num_base_bdevs_operational": 3, 00:13:04.996 "process": { 00:13:04.996 "type": "rebuild", 00:13:04.996 "target": "spare", 00:13:04.996 "progress": { 00:13:04.996 "blocks": 18432, 00:13:04.996 "percent": 28 00:13:04.996 } 00:13:04.996 }, 00:13:04.996 "base_bdevs_list": [ 00:13:04.996 { 00:13:04.996 "name": "spare", 00:13:04.996 "uuid": "3fe1edd0-f1a2-57df-9e90-ad085859747f", 00:13:04.996 "is_configured": true, 00:13:04.996 "data_offset": 0, 00:13:04.996 "data_size": 65536 00:13:04.996 }, 00:13:04.996 { 00:13:04.996 "name": null, 00:13:04.996 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:04.996 "is_configured": false, 00:13:04.996 "data_offset": 0, 00:13:04.996 "data_size": 65536 00:13:04.996 }, 00:13:04.996 { 00:13:04.996 "name": "BaseBdev3", 00:13:04.996 "uuid": "09878590-9963-5efd-800a-1862515421c6", 00:13:04.996 "is_configured": true, 00:13:04.996 "data_offset": 0, 00:13:04.996 "data_size": 65536 00:13:04.996 }, 00:13:04.996 { 00:13:04.996 "name": "BaseBdev4", 00:13:04.996 "uuid": "6bfd443d-22ee-5db5-8f9b-d054694b31be", 00:13:04.996 "is_configured": true, 00:13:04.996 "data_offset": 0, 00:13:04.996 "data_size": 65536 00:13:04.996 } 00:13:04.996 ] 00:13:04.996 }' 00:13:04.996 15:17:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:05.256 15:17:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:05.256 15:17:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:05.256 [2024-11-27 15:17:33.120884] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:13:05.256 [2024-11-27 15:17:33.121478] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:13:05.256 15:17:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:05.256 15:17:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@706 -- # local timeout=394 00:13:05.256 15:17:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:05.256 15:17:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:05.256 15:17:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:05.256 15:17:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:05.256 15:17:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:05.256 15:17:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:05.256 15:17:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:05.256 15:17:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:05.256 15:17:33 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:05.256 15:17:33 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:05.256 15:17:33 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:05.256 15:17:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:05.256 "name": "raid_bdev1", 00:13:05.256 "uuid": "bfafadfd-f3ca-409e-b748-7b5977cf95fe", 00:13:05.256 "strip_size_kb": 0, 00:13:05.256 "state": "online", 00:13:05.256 "raid_level": "raid1", 00:13:05.256 "superblock": false, 00:13:05.256 "num_base_bdevs": 4, 00:13:05.256 "num_base_bdevs_discovered": 3, 00:13:05.256 "num_base_bdevs_operational": 3, 00:13:05.256 "process": { 00:13:05.256 "type": "rebuild", 00:13:05.256 "target": "spare", 00:13:05.256 "progress": { 00:13:05.256 "blocks": 20480, 00:13:05.256 "percent": 31 00:13:05.256 } 00:13:05.256 }, 00:13:05.256 "base_bdevs_list": [ 00:13:05.256 { 00:13:05.256 "name": "spare", 00:13:05.256 "uuid": "3fe1edd0-f1a2-57df-9e90-ad085859747f", 00:13:05.256 "is_configured": true, 00:13:05.256 "data_offset": 0, 00:13:05.256 "data_size": 65536 00:13:05.256 }, 00:13:05.256 { 00:13:05.256 "name": null, 00:13:05.256 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:05.256 "is_configured": false, 00:13:05.256 "data_offset": 0, 00:13:05.256 "data_size": 65536 00:13:05.256 }, 00:13:05.256 { 00:13:05.256 "name": "BaseBdev3", 00:13:05.256 "uuid": "09878590-9963-5efd-800a-1862515421c6", 00:13:05.256 "is_configured": true, 00:13:05.256 "data_offset": 0, 00:13:05.256 "data_size": 65536 00:13:05.256 }, 00:13:05.256 { 00:13:05.256 "name": "BaseBdev4", 00:13:05.256 "uuid": "6bfd443d-22ee-5db5-8f9b-d054694b31be", 00:13:05.256 "is_configured": true, 00:13:05.256 "data_offset": 0, 00:13:05.256 "data_size": 65536 00:13:05.256 } 00:13:05.256 ] 00:13:05.256 }' 00:13:05.256 15:17:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:05.256 15:17:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:05.256 15:17:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:05.256 [2024-11-27 15:17:33.244137] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:13:05.256 15:17:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:05.256 15:17:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:05.826 145.50 IOPS, 436.50 MiB/s [2024-11-27T15:17:33.933Z] [2024-11-27 15:17:33.912161] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:13:06.397 [2024-11-27 15:17:34.277035] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 38912 offset_begin: 36864 offset_end: 43008 00:13:06.397 15:17:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:06.397 15:17:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:06.397 15:17:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:06.397 15:17:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:06.397 15:17:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:06.397 15:17:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:06.397 123.00 IOPS, 369.00 MiB/s [2024-11-27T15:17:34.504Z] 15:17:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:06.397 15:17:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:06.397 15:17:34 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:06.397 15:17:34 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:06.397 15:17:34 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:06.397 15:17:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:06.397 "name": "raid_bdev1", 00:13:06.397 "uuid": "bfafadfd-f3ca-409e-b748-7b5977cf95fe", 00:13:06.397 "strip_size_kb": 0, 00:13:06.397 "state": "online", 00:13:06.397 "raid_level": "raid1", 00:13:06.397 "superblock": false, 00:13:06.397 "num_base_bdevs": 4, 00:13:06.397 "num_base_bdevs_discovered": 3, 00:13:06.397 "num_base_bdevs_operational": 3, 00:13:06.397 "process": { 00:13:06.397 "type": "rebuild", 00:13:06.397 "target": "spare", 00:13:06.397 "progress": { 00:13:06.397 "blocks": 38912, 00:13:06.397 "percent": 59 00:13:06.397 } 00:13:06.397 }, 00:13:06.397 "base_bdevs_list": [ 00:13:06.397 { 00:13:06.397 "name": "spare", 00:13:06.397 "uuid": "3fe1edd0-f1a2-57df-9e90-ad085859747f", 00:13:06.397 "is_configured": true, 00:13:06.397 "data_offset": 0, 00:13:06.397 "data_size": 65536 00:13:06.397 }, 00:13:06.397 { 00:13:06.397 "name": null, 00:13:06.397 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:06.397 "is_configured": false, 00:13:06.397 "data_offset": 0, 00:13:06.397 "data_size": 65536 00:13:06.397 }, 00:13:06.397 { 00:13:06.397 "name": "BaseBdev3", 00:13:06.397 "uuid": "09878590-9963-5efd-800a-1862515421c6", 00:13:06.397 "is_configured": true, 00:13:06.397 "data_offset": 0, 00:13:06.397 "data_size": 65536 00:13:06.397 }, 00:13:06.397 { 00:13:06.397 "name": "BaseBdev4", 00:13:06.397 "uuid": "6bfd443d-22ee-5db5-8f9b-d054694b31be", 00:13:06.397 "is_configured": true, 00:13:06.397 "data_offset": 0, 00:13:06.397 "data_size": 65536 00:13:06.397 } 00:13:06.397 ] 00:13:06.397 }' 00:13:06.397 15:17:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:06.397 15:17:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:06.397 15:17:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:06.397 15:17:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:06.397 15:17:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:06.397 [2024-11-27 15:17:34.482145] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 40960 offset_begin: 36864 offset_end: 43008 00:13:06.966 [2024-11-27 15:17:34.913694] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 47104 offset_begin: 43008 offset_end: 49152 00:13:06.966 [2024-11-27 15:17:34.914218] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 47104 offset_begin: 43008 offset_end: 49152 00:13:07.225 [2024-11-27 15:17:35.247426] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 51200 offset_begin: 49152 offset_end: 55296 00:13:07.486 110.17 IOPS, 330.50 MiB/s [2024-11-27T15:17:35.593Z] 15:17:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:07.486 15:17:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:07.486 15:17:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:07.486 15:17:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:07.486 15:17:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:07.486 15:17:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:07.486 15:17:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:07.486 15:17:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:07.486 15:17:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:07.486 15:17:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:07.486 [2024-11-27 15:17:35.456664] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 53248 offset_begin: 49152 offset_end: 55296 00:13:07.486 [2024-11-27 15:17:35.457302] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 53248 offset_begin: 49152 offset_end: 55296 00:13:07.486 15:17:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:07.486 15:17:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:07.486 "name": "raid_bdev1", 00:13:07.486 "uuid": "bfafadfd-f3ca-409e-b748-7b5977cf95fe", 00:13:07.486 "strip_size_kb": 0, 00:13:07.486 "state": "online", 00:13:07.486 "raid_level": "raid1", 00:13:07.486 "superblock": false, 00:13:07.486 "num_base_bdevs": 4, 00:13:07.486 "num_base_bdevs_discovered": 3, 00:13:07.486 "num_base_bdevs_operational": 3, 00:13:07.486 "process": { 00:13:07.486 "type": "rebuild", 00:13:07.486 "target": "spare", 00:13:07.486 "progress": { 00:13:07.486 "blocks": 51200, 00:13:07.486 "percent": 78 00:13:07.486 } 00:13:07.486 }, 00:13:07.486 "base_bdevs_list": [ 00:13:07.486 { 00:13:07.486 "name": "spare", 00:13:07.486 "uuid": "3fe1edd0-f1a2-57df-9e90-ad085859747f", 00:13:07.486 "is_configured": true, 00:13:07.486 "data_offset": 0, 00:13:07.486 "data_size": 65536 00:13:07.486 }, 00:13:07.486 { 00:13:07.486 "name": null, 00:13:07.486 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:07.486 "is_configured": false, 00:13:07.486 "data_offset": 0, 00:13:07.486 "data_size": 65536 00:13:07.486 }, 00:13:07.486 { 00:13:07.486 "name": "BaseBdev3", 00:13:07.486 "uuid": "09878590-9963-5efd-800a-1862515421c6", 00:13:07.486 "is_configured": true, 00:13:07.486 "data_offset": 0, 00:13:07.486 "data_size": 65536 00:13:07.486 }, 00:13:07.486 { 00:13:07.486 "name": "BaseBdev4", 00:13:07.486 "uuid": "6bfd443d-22ee-5db5-8f9b-d054694b31be", 00:13:07.486 "is_configured": true, 00:13:07.486 "data_offset": 0, 00:13:07.486 "data_size": 65536 00:13:07.486 } 00:13:07.486 ] 00:13:07.486 }' 00:13:07.486 15:17:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:07.486 15:17:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:07.486 15:17:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:07.746 15:17:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:07.746 15:17:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:07.746 [2024-11-27 15:17:35.780293] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 57344 offset_begin: 55296 offset_end: 61440 00:13:07.746 [2024-11-27 15:17:35.781367] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 57344 offset_begin: 55296 offset_end: 61440 00:13:08.315 99.14 IOPS, 297.43 MiB/s [2024-11-27T15:17:36.422Z] [2024-11-27 15:17:36.317486] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:13:08.575 [2024-11-27 15:17:36.422651] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:13:08.575 [2024-11-27 15:17:36.426081] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:08.575 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:08.575 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:08.575 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:08.575 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:08.575 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:08.575 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:08.575 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:08.575 15:17:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:08.575 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:08.575 15:17:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:08.575 15:17:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:08.575 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:08.575 "name": "raid_bdev1", 00:13:08.575 "uuid": "bfafadfd-f3ca-409e-b748-7b5977cf95fe", 00:13:08.575 "strip_size_kb": 0, 00:13:08.575 "state": "online", 00:13:08.575 "raid_level": "raid1", 00:13:08.575 "superblock": false, 00:13:08.575 "num_base_bdevs": 4, 00:13:08.575 "num_base_bdevs_discovered": 3, 00:13:08.575 "num_base_bdevs_operational": 3, 00:13:08.575 "base_bdevs_list": [ 00:13:08.575 { 00:13:08.575 "name": "spare", 00:13:08.575 "uuid": "3fe1edd0-f1a2-57df-9e90-ad085859747f", 00:13:08.575 "is_configured": true, 00:13:08.575 "data_offset": 0, 00:13:08.575 "data_size": 65536 00:13:08.575 }, 00:13:08.575 { 00:13:08.575 "name": null, 00:13:08.575 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:08.575 "is_configured": false, 00:13:08.575 "data_offset": 0, 00:13:08.575 "data_size": 65536 00:13:08.575 }, 00:13:08.575 { 00:13:08.575 "name": "BaseBdev3", 00:13:08.575 "uuid": "09878590-9963-5efd-800a-1862515421c6", 00:13:08.575 "is_configured": true, 00:13:08.575 "data_offset": 0, 00:13:08.575 "data_size": 65536 00:13:08.575 }, 00:13:08.575 { 00:13:08.575 "name": "BaseBdev4", 00:13:08.575 "uuid": "6bfd443d-22ee-5db5-8f9b-d054694b31be", 00:13:08.575 "is_configured": true, 00:13:08.575 "data_offset": 0, 00:13:08.575 "data_size": 65536 00:13:08.575 } 00:13:08.575 ] 00:13:08.575 }' 00:13:08.575 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:08.575 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@709 -- # break 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:08.835 "name": "raid_bdev1", 00:13:08.835 "uuid": "bfafadfd-f3ca-409e-b748-7b5977cf95fe", 00:13:08.835 "strip_size_kb": 0, 00:13:08.835 "state": "online", 00:13:08.835 "raid_level": "raid1", 00:13:08.835 "superblock": false, 00:13:08.835 "num_base_bdevs": 4, 00:13:08.835 "num_base_bdevs_discovered": 3, 00:13:08.835 "num_base_bdevs_operational": 3, 00:13:08.835 "base_bdevs_list": [ 00:13:08.835 { 00:13:08.835 "name": "spare", 00:13:08.835 "uuid": "3fe1edd0-f1a2-57df-9e90-ad085859747f", 00:13:08.835 "is_configured": true, 00:13:08.835 "data_offset": 0, 00:13:08.835 "data_size": 65536 00:13:08.835 }, 00:13:08.835 { 00:13:08.835 "name": null, 00:13:08.835 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:08.835 "is_configured": false, 00:13:08.835 "data_offset": 0, 00:13:08.835 "data_size": 65536 00:13:08.835 }, 00:13:08.835 { 00:13:08.835 "name": "BaseBdev3", 00:13:08.835 "uuid": "09878590-9963-5efd-800a-1862515421c6", 00:13:08.835 "is_configured": true, 00:13:08.835 "data_offset": 0, 00:13:08.835 "data_size": 65536 00:13:08.835 }, 00:13:08.835 { 00:13:08.835 "name": "BaseBdev4", 00:13:08.835 "uuid": "6bfd443d-22ee-5db5-8f9b-d054694b31be", 00:13:08.835 "is_configured": true, 00:13:08.835 "data_offset": 0, 00:13:08.835 "data_size": 65536 00:13:08.835 } 00:13:08.835 ] 00:13:08.835 }' 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:08.835 "name": "raid_bdev1", 00:13:08.835 "uuid": "bfafadfd-f3ca-409e-b748-7b5977cf95fe", 00:13:08.835 "strip_size_kb": 0, 00:13:08.835 "state": "online", 00:13:08.835 "raid_level": "raid1", 00:13:08.835 "superblock": false, 00:13:08.835 "num_base_bdevs": 4, 00:13:08.835 "num_base_bdevs_discovered": 3, 00:13:08.835 "num_base_bdevs_operational": 3, 00:13:08.835 "base_bdevs_list": [ 00:13:08.835 { 00:13:08.835 "name": "spare", 00:13:08.835 "uuid": "3fe1edd0-f1a2-57df-9e90-ad085859747f", 00:13:08.835 "is_configured": true, 00:13:08.835 "data_offset": 0, 00:13:08.835 "data_size": 65536 00:13:08.835 }, 00:13:08.835 { 00:13:08.835 "name": null, 00:13:08.835 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:08.835 "is_configured": false, 00:13:08.835 "data_offset": 0, 00:13:08.835 "data_size": 65536 00:13:08.835 }, 00:13:08.835 { 00:13:08.835 "name": "BaseBdev3", 00:13:08.835 "uuid": "09878590-9963-5efd-800a-1862515421c6", 00:13:08.835 "is_configured": true, 00:13:08.835 "data_offset": 0, 00:13:08.835 "data_size": 65536 00:13:08.835 }, 00:13:08.835 { 00:13:08.835 "name": "BaseBdev4", 00:13:08.835 "uuid": "6bfd443d-22ee-5db5-8f9b-d054694b31be", 00:13:08.835 "is_configured": true, 00:13:08.835 "data_offset": 0, 00:13:08.835 "data_size": 65536 00:13:08.835 } 00:13:08.835 ] 00:13:08.835 }' 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:08.835 15:17:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:09.405 89.50 IOPS, 268.50 MiB/s [2024-11-27T15:17:37.512Z] 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:09.405 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:09.405 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:09.405 [2024-11-27 15:17:37.315413] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:09.405 [2024-11-27 15:17:37.315512] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:09.405 00:13:09.405 Latency(us) 00:13:09.405 [2024-11-27T15:17:37.512Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:09.405 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:13:09.405 raid_bdev1 : 8.12 88.54 265.63 0.00 0.00 16209.57 300.49 119052.30 00:13:09.405 [2024-11-27T15:17:37.512Z] =================================================================================================================== 00:13:09.405 [2024-11-27T15:17:37.512Z] Total : 88.54 265.63 0.00 0.00 16209.57 300.49 119052.30 00:13:09.405 [2024-11-27 15:17:37.394530] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:09.405 [2024-11-27 15:17:37.394622] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:09.405 [2024-11-27 15:17:37.394739] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:09.405 [2024-11-27 15:17:37.394787] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:13:09.405 { 00:13:09.405 "results": [ 00:13:09.405 { 00:13:09.405 "job": "raid_bdev1", 00:13:09.405 "core_mask": "0x1", 00:13:09.405 "workload": "randrw", 00:13:09.405 "percentage": 50, 00:13:09.405 "status": "finished", 00:13:09.405 "queue_depth": 2, 00:13:09.405 "io_size": 3145728, 00:13:09.405 "runtime": 8.120331, 00:13:09.405 "iops": 88.54318869514063, 00:13:09.405 "mibps": 265.62956608542186, 00:13:09.405 "io_failed": 0, 00:13:09.405 "io_timeout": 0, 00:13:09.405 "avg_latency_us": 16209.57091545147, 00:13:09.405 "min_latency_us": 300.49257641921395, 00:13:09.405 "max_latency_us": 119052.29694323144 00:13:09.405 } 00:13:09.405 ], 00:13:09.405 "core_count": 1 00:13:09.405 } 00:13:09.405 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:09.405 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:09.405 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # jq length 00:13:09.405 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:09.405 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:09.405 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:09.405 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:13:09.405 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:13:09.405 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:13:09.405 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:13:09.405 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:09.405 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:13:09.405 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:09.405 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:13:09.405 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:09.405 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:13:09.405 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:09.405 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:09.405 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:13:09.682 /dev/nbd0 00:13:09.682 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:09.682 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:09.682 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:13:09.682 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:13:09.682 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:09.682 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:09.682 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:13:09.682 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:13:09.682 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:09.682 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:09.682 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:09.682 1+0 records in 00:13:09.682 1+0 records out 00:13:09.682 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000527982 s, 7.8 MB/s 00:13:09.682 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:09.682 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:13:09.683 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:09.683 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:09.683 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:13:09.683 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:09.683 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:09.683 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:13:09.683 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z '' ']' 00:13:09.683 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@728 -- # continue 00:13:09.683 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:13:09.683 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev3 ']' 00:13:09.683 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev3 /dev/nbd1 00:13:09.683 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:09.683 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev3') 00:13:09.683 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:09.683 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:13:09.683 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:09.683 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:13:09.683 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:09.683 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:09.683 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev3 /dev/nbd1 00:13:09.961 /dev/nbd1 00:13:09.961 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:09.961 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:09.961 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:13:09.961 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:13:09.961 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:09.961 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:09.961 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:13:09.961 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:13:09.961 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:09.961 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:09.961 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:09.961 1+0 records in 00:13:09.961 1+0 records out 00:13:09.961 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000287602 s, 14.2 MB/s 00:13:09.961 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:09.961 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:13:09.961 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:09.961 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:09.961 15:17:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:13:09.961 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:09.961 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:09.961 15:17:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:13:09.961 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:13:09.961 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:09.961 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:13:09.961 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:09.961 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:13:09.961 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:09.961 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:10.220 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:10.220 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:10.220 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:10.220 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:10.220 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:10.220 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:10.220 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:13:10.220 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:10.220 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:13:10.220 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev4 ']' 00:13:10.220 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev4 /dev/nbd1 00:13:10.220 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:10.220 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev4') 00:13:10.220 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:10.220 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:13:10.220 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:10.220 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:13:10.220 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:10.220 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:10.220 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev4 /dev/nbd1 00:13:10.480 /dev/nbd1 00:13:10.480 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:10.480 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:10.480 15:17:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:13:10.480 15:17:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:13:10.480 15:17:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:10.480 15:17:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:10.480 15:17:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:13:10.480 15:17:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:13:10.480 15:17:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:10.480 15:17:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:10.480 15:17:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:10.480 1+0 records in 00:13:10.480 1+0 records out 00:13:10.480 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000522529 s, 7.8 MB/s 00:13:10.480 15:17:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:10.480 15:17:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:13:10.480 15:17:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:10.480 15:17:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:10.480 15:17:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:13:10.480 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:10.480 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:10.480 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:13:10.480 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:13:10.480 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:10.480 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:13:10.480 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:10.480 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:13:10.480 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:10.480 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:10.739 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:10.739 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:10.739 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:10.739 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:10.739 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:10.739 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:10.739 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:13:10.739 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:10.739 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:13:10.739 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:10.739 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:13:10.739 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:10.739 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:13:10.739 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:10.739 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:10.999 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:10.999 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:10.999 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:10.999 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:10.999 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:10.999 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:10.999 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:13:10.999 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:10.999 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:13:10.999 15:17:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@784 -- # killprocess 89530 00:13:10.999 15:17:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@954 -- # '[' -z 89530 ']' 00:13:10.999 15:17:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@958 -- # kill -0 89530 00:13:10.999 15:17:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@959 -- # uname 00:13:10.999 15:17:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:10.999 15:17:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 89530 00:13:10.999 15:17:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:10.999 15:17:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:10.999 15:17:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@972 -- # echo 'killing process with pid 89530' 00:13:10.999 killing process with pid 89530 00:13:10.999 Received shutdown signal, test time was about 9.704145 seconds 00:13:10.999 00:13:10.999 Latency(us) 00:13:10.999 [2024-11-27T15:17:39.106Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:10.999 [2024-11-27T15:17:39.106Z] =================================================================================================================== 00:13:10.999 [2024-11-27T15:17:39.106Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:13:10.999 15:17:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@973 -- # kill 89530 00:13:10.999 [2024-11-27 15:17:38.971973] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:10.999 15:17:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@978 -- # wait 89530 00:13:10.999 [2024-11-27 15:17:39.019537] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:11.259 ************************************ 00:13:11.259 END TEST raid_rebuild_test_io 00:13:11.259 ************************************ 00:13:11.259 15:17:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@786 -- # return 0 00:13:11.259 00:13:11.259 real 0m11.671s 00:13:11.259 user 0m15.090s 00:13:11.259 sys 0m1.758s 00:13:11.259 15:17:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:11.259 15:17:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:11.259 15:17:39 bdev_raid -- bdev/bdev_raid.sh@981 -- # run_test raid_rebuild_test_sb_io raid_rebuild_test raid1 4 true true true 00:13:11.259 15:17:39 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:13:11.259 15:17:39 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:11.259 15:17:39 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:11.259 ************************************ 00:13:11.259 START TEST raid_rebuild_test_sb_io 00:13:11.259 ************************************ 00:13:11.259 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 4 true true true 00:13:11.259 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:13:11.259 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:13:11.259 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:13:11.259 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:13:11.259 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:13:11.259 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:13:11.259 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:11.259 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:13:11.259 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:11.259 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:11.259 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:13:11.259 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:11.259 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:11.259 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:13:11.259 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:11.259 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:11.259 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:13:11.259 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:11.259 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:11.259 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:13:11.259 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:13:11.259 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:13:11.259 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:13:11.259 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:13:11.259 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:13:11.259 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:13:11.260 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:13:11.260 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:13:11.260 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:13:11.260 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:13:11.260 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@597 -- # raid_pid=89929 00:13:11.260 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:13:11.260 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 89929 00:13:11.260 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@835 -- # '[' -z 89929 ']' 00:13:11.260 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:11.260 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@840 -- # local max_retries=100 00:13:11.260 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:11.260 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:11.260 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@844 -- # xtrace_disable 00:13:11.260 15:17:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:11.519 [2024-11-27 15:17:39.401976] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:13:11.519 [2024-11-27 15:17:39.402179] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid89929 ] 00:13:11.519 I/O size of 3145728 is greater than zero copy threshold (65536). 00:13:11.519 Zero copy mechanism will not be used. 00:13:11.519 [2024-11-27 15:17:39.569292] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:11.519 [2024-11-27 15:17:39.594665] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:11.778 [2024-11-27 15:17:39.638176] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:11.778 [2024-11-27 15:17:39.638303] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:12.347 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:13:12.347 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # return 0 00:13:12.347 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:12.347 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:13:12.347 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.347 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:12.347 BaseBdev1_malloc 00:13:12.347 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.347 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:12.347 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.347 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:12.347 [2024-11-27 15:17:40.266812] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:12.347 [2024-11-27 15:17:40.266952] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:12.347 [2024-11-27 15:17:40.267015] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:13:12.347 [2024-11-27 15:17:40.267058] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:12.347 [2024-11-27 15:17:40.269196] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:12.347 [2024-11-27 15:17:40.269291] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:12.347 BaseBdev1 00:13:12.347 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.347 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:12.347 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:13:12.347 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.347 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:12.347 BaseBdev2_malloc 00:13:12.347 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.347 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:13:12.347 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.347 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:12.347 [2024-11-27 15:17:40.295974] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:13:12.347 [2024-11-27 15:17:40.296074] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:12.347 [2024-11-27 15:17:40.296115] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:13:12.347 [2024-11-27 15:17:40.296126] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:12.348 [2024-11-27 15:17:40.298236] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:12.348 [2024-11-27 15:17:40.298304] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:13:12.348 BaseBdev2 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:12.348 BaseBdev3_malloc 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:12.348 [2024-11-27 15:17:40.324752] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:13:12.348 [2024-11-27 15:17:40.324866] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:12.348 [2024-11-27 15:17:40.324905] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:13:12.348 [2024-11-27 15:17:40.324962] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:12.348 [2024-11-27 15:17:40.326959] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:12.348 [2024-11-27 15:17:40.327024] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:13:12.348 BaseBdev3 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:12.348 BaseBdev4_malloc 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:12.348 [2024-11-27 15:17:40.370888] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:13:12.348 [2024-11-27 15:17:40.371045] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:12.348 [2024-11-27 15:17:40.371116] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:13:12.348 [2024-11-27 15:17:40.371170] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:12.348 [2024-11-27 15:17:40.374419] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:12.348 [2024-11-27 15:17:40.374523] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:13:12.348 BaseBdev4 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:12.348 spare_malloc 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:12.348 spare_delay 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:12.348 [2024-11-27 15:17:40.412790] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:12.348 [2024-11-27 15:17:40.412903] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:12.348 [2024-11-27 15:17:40.412938] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:13:12.348 [2024-11-27 15:17:40.412947] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:12.348 [2024-11-27 15:17:40.415035] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:12.348 [2024-11-27 15:17:40.415104] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:12.348 spare 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:12.348 [2024-11-27 15:17:40.424848] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:12.348 [2024-11-27 15:17:40.426648] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:12.348 [2024-11-27 15:17:40.426762] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:12.348 [2024-11-27 15:17:40.426837] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:13:12.348 [2024-11-27 15:17:40.427055] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:13:12.348 [2024-11-27 15:17:40.427106] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:13:12.348 [2024-11-27 15:17:40.427383] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:13:12.348 [2024-11-27 15:17:40.427600] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:13:12.348 [2024-11-27 15:17:40.427658] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:13:12.348 [2024-11-27 15:17:40.427804] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:12.348 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:12.607 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.607 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:12.607 "name": "raid_bdev1", 00:13:12.607 "uuid": "0e636b34-1fee-4bd2-bbb1-1f06f968ec67", 00:13:12.607 "strip_size_kb": 0, 00:13:12.607 "state": "online", 00:13:12.607 "raid_level": "raid1", 00:13:12.607 "superblock": true, 00:13:12.607 "num_base_bdevs": 4, 00:13:12.607 "num_base_bdevs_discovered": 4, 00:13:12.607 "num_base_bdevs_operational": 4, 00:13:12.607 "base_bdevs_list": [ 00:13:12.607 { 00:13:12.607 "name": "BaseBdev1", 00:13:12.607 "uuid": "bb978c91-7f5a-5bfd-ad6d-05a8e610648a", 00:13:12.607 "is_configured": true, 00:13:12.607 "data_offset": 2048, 00:13:12.607 "data_size": 63488 00:13:12.607 }, 00:13:12.607 { 00:13:12.607 "name": "BaseBdev2", 00:13:12.607 "uuid": "3d99c079-b40c-5fd1-9da2-35620d3e4cee", 00:13:12.607 "is_configured": true, 00:13:12.607 "data_offset": 2048, 00:13:12.607 "data_size": 63488 00:13:12.607 }, 00:13:12.607 { 00:13:12.607 "name": "BaseBdev3", 00:13:12.607 "uuid": "36988339-b583-5c5c-974f-fd338828ec4f", 00:13:12.607 "is_configured": true, 00:13:12.607 "data_offset": 2048, 00:13:12.607 "data_size": 63488 00:13:12.607 }, 00:13:12.607 { 00:13:12.607 "name": "BaseBdev4", 00:13:12.607 "uuid": "57e9c1d7-cb24-59ed-9066-0d5f0ec96dab", 00:13:12.607 "is_configured": true, 00:13:12.607 "data_offset": 2048, 00:13:12.607 "data_size": 63488 00:13:12.607 } 00:13:12.607 ] 00:13:12.607 }' 00:13:12.607 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:12.607 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:12.866 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:13:12.866 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:12.866 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.866 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:12.866 [2024-11-27 15:17:40.928264] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:12.866 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.866 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:13:12.866 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:12.866 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.866 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:12.867 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:13:13.127 15:17:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:13.127 15:17:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:13:13.127 15:17:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:13:13.127 15:17:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:13:13.127 15:17:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:13:13.127 15:17:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:13.127 15:17:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:13.127 [2024-11-27 15:17:41.015829] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:13.127 15:17:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:13.127 15:17:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:13.127 15:17:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:13.127 15:17:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:13.127 15:17:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:13.127 15:17:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:13.127 15:17:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:13.127 15:17:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:13.127 15:17:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:13.127 15:17:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:13.127 15:17:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:13.127 15:17:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:13.127 15:17:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:13.127 15:17:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:13.127 15:17:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:13.127 15:17:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:13.127 15:17:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:13.127 "name": "raid_bdev1", 00:13:13.127 "uuid": "0e636b34-1fee-4bd2-bbb1-1f06f968ec67", 00:13:13.127 "strip_size_kb": 0, 00:13:13.127 "state": "online", 00:13:13.127 "raid_level": "raid1", 00:13:13.127 "superblock": true, 00:13:13.127 "num_base_bdevs": 4, 00:13:13.127 "num_base_bdevs_discovered": 3, 00:13:13.127 "num_base_bdevs_operational": 3, 00:13:13.127 "base_bdevs_list": [ 00:13:13.127 { 00:13:13.127 "name": null, 00:13:13.127 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:13.127 "is_configured": false, 00:13:13.127 "data_offset": 0, 00:13:13.127 "data_size": 63488 00:13:13.127 }, 00:13:13.127 { 00:13:13.127 "name": "BaseBdev2", 00:13:13.127 "uuid": "3d99c079-b40c-5fd1-9da2-35620d3e4cee", 00:13:13.127 "is_configured": true, 00:13:13.127 "data_offset": 2048, 00:13:13.127 "data_size": 63488 00:13:13.127 }, 00:13:13.127 { 00:13:13.127 "name": "BaseBdev3", 00:13:13.127 "uuid": "36988339-b583-5c5c-974f-fd338828ec4f", 00:13:13.127 "is_configured": true, 00:13:13.127 "data_offset": 2048, 00:13:13.127 "data_size": 63488 00:13:13.127 }, 00:13:13.127 { 00:13:13.127 "name": "BaseBdev4", 00:13:13.127 "uuid": "57e9c1d7-cb24-59ed-9066-0d5f0ec96dab", 00:13:13.127 "is_configured": true, 00:13:13.127 "data_offset": 2048, 00:13:13.127 "data_size": 63488 00:13:13.127 } 00:13:13.127 ] 00:13:13.127 }' 00:13:13.127 15:17:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:13.127 15:17:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:13.127 [2024-11-27 15:17:41.109744] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:13:13.127 I/O size of 3145728 is greater than zero copy threshold (65536). 00:13:13.127 Zero copy mechanism will not be used. 00:13:13.127 Running I/O for 60 seconds... 00:13:13.388 15:17:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:13.388 15:17:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:13.388 15:17:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:13.388 [2024-11-27 15:17:41.460247] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:13.388 15:17:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:13.388 15:17:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:13:13.647 [2024-11-27 15:17:41.516458] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:13:13.647 [2024-11-27 15:17:41.518446] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:13.647 [2024-11-27 15:17:41.654756] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:13.907 [2024-11-27 15:17:41.863442] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:13.907 [2024-11-27 15:17:41.863878] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:14.426 204.00 IOPS, 612.00 MiB/s [2024-11-27T15:17:42.533Z] 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:14.426 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:14.426 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:14.426 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:14.426 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:14.426 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:14.426 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:14.426 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:14.426 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:14.685 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:14.685 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:14.685 "name": "raid_bdev1", 00:13:14.685 "uuid": "0e636b34-1fee-4bd2-bbb1-1f06f968ec67", 00:13:14.685 "strip_size_kb": 0, 00:13:14.685 "state": "online", 00:13:14.685 "raid_level": "raid1", 00:13:14.685 "superblock": true, 00:13:14.685 "num_base_bdevs": 4, 00:13:14.685 "num_base_bdevs_discovered": 4, 00:13:14.685 "num_base_bdevs_operational": 4, 00:13:14.685 "process": { 00:13:14.685 "type": "rebuild", 00:13:14.685 "target": "spare", 00:13:14.685 "progress": { 00:13:14.685 "blocks": 12288, 00:13:14.685 "percent": 19 00:13:14.685 } 00:13:14.685 }, 00:13:14.685 "base_bdevs_list": [ 00:13:14.686 { 00:13:14.686 "name": "spare", 00:13:14.686 "uuid": "678e143c-d959-5d2f-9795-cd1e0ad776f6", 00:13:14.686 "is_configured": true, 00:13:14.686 "data_offset": 2048, 00:13:14.686 "data_size": 63488 00:13:14.686 }, 00:13:14.686 { 00:13:14.686 "name": "BaseBdev2", 00:13:14.686 "uuid": "3d99c079-b40c-5fd1-9da2-35620d3e4cee", 00:13:14.686 "is_configured": true, 00:13:14.686 "data_offset": 2048, 00:13:14.686 "data_size": 63488 00:13:14.686 }, 00:13:14.686 { 00:13:14.686 "name": "BaseBdev3", 00:13:14.686 "uuid": "36988339-b583-5c5c-974f-fd338828ec4f", 00:13:14.686 "is_configured": true, 00:13:14.686 "data_offset": 2048, 00:13:14.686 "data_size": 63488 00:13:14.686 }, 00:13:14.686 { 00:13:14.686 "name": "BaseBdev4", 00:13:14.686 "uuid": "57e9c1d7-cb24-59ed-9066-0d5f0ec96dab", 00:13:14.686 "is_configured": true, 00:13:14.686 "data_offset": 2048, 00:13:14.686 "data_size": 63488 00:13:14.686 } 00:13:14.686 ] 00:13:14.686 }' 00:13:14.686 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:14.686 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:14.686 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:14.686 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:14.686 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:14.686 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:14.686 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:14.686 [2024-11-27 15:17:42.654093] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:14.686 [2024-11-27 15:17:42.667696] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:13:14.686 [2024-11-27 15:17:42.726853] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:14.686 [2024-11-27 15:17:42.743178] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:14.686 [2024-11-27 15:17:42.743290] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:14.686 [2024-11-27 15:17:42.743320] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:14.686 [2024-11-27 15:17:42.754375] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000006080 00:13:14.686 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:14.686 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:14.686 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:14.686 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:14.686 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:14.686 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:14.686 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:14.686 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:14.686 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:14.686 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:14.686 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:14.686 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:14.686 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:14.686 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:14.686 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:14.946 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:14.946 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:14.946 "name": "raid_bdev1", 00:13:14.946 "uuid": "0e636b34-1fee-4bd2-bbb1-1f06f968ec67", 00:13:14.946 "strip_size_kb": 0, 00:13:14.946 "state": "online", 00:13:14.946 "raid_level": "raid1", 00:13:14.946 "superblock": true, 00:13:14.946 "num_base_bdevs": 4, 00:13:14.946 "num_base_bdevs_discovered": 3, 00:13:14.946 "num_base_bdevs_operational": 3, 00:13:14.946 "base_bdevs_list": [ 00:13:14.946 { 00:13:14.946 "name": null, 00:13:14.946 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:14.946 "is_configured": false, 00:13:14.946 "data_offset": 0, 00:13:14.946 "data_size": 63488 00:13:14.946 }, 00:13:14.946 { 00:13:14.946 "name": "BaseBdev2", 00:13:14.946 "uuid": "3d99c079-b40c-5fd1-9da2-35620d3e4cee", 00:13:14.946 "is_configured": true, 00:13:14.946 "data_offset": 2048, 00:13:14.946 "data_size": 63488 00:13:14.946 }, 00:13:14.946 { 00:13:14.946 "name": "BaseBdev3", 00:13:14.946 "uuid": "36988339-b583-5c5c-974f-fd338828ec4f", 00:13:14.946 "is_configured": true, 00:13:14.946 "data_offset": 2048, 00:13:14.946 "data_size": 63488 00:13:14.946 }, 00:13:14.946 { 00:13:14.946 "name": "BaseBdev4", 00:13:14.946 "uuid": "57e9c1d7-cb24-59ed-9066-0d5f0ec96dab", 00:13:14.946 "is_configured": true, 00:13:14.946 "data_offset": 2048, 00:13:14.946 "data_size": 63488 00:13:14.946 } 00:13:14.946 ] 00:13:14.946 }' 00:13:14.946 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:14.946 15:17:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:15.206 185.00 IOPS, 555.00 MiB/s [2024-11-27T15:17:43.313Z] 15:17:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:15.206 15:17:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:15.206 15:17:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:15.206 15:17:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:15.206 15:17:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:15.206 15:17:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:15.206 15:17:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:15.206 15:17:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:15.206 15:17:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:15.206 15:17:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:15.206 15:17:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:15.206 "name": "raid_bdev1", 00:13:15.206 "uuid": "0e636b34-1fee-4bd2-bbb1-1f06f968ec67", 00:13:15.206 "strip_size_kb": 0, 00:13:15.206 "state": "online", 00:13:15.206 "raid_level": "raid1", 00:13:15.206 "superblock": true, 00:13:15.206 "num_base_bdevs": 4, 00:13:15.206 "num_base_bdevs_discovered": 3, 00:13:15.206 "num_base_bdevs_operational": 3, 00:13:15.206 "base_bdevs_list": [ 00:13:15.206 { 00:13:15.206 "name": null, 00:13:15.206 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:15.206 "is_configured": false, 00:13:15.206 "data_offset": 0, 00:13:15.206 "data_size": 63488 00:13:15.206 }, 00:13:15.206 { 00:13:15.206 "name": "BaseBdev2", 00:13:15.206 "uuid": "3d99c079-b40c-5fd1-9da2-35620d3e4cee", 00:13:15.206 "is_configured": true, 00:13:15.206 "data_offset": 2048, 00:13:15.206 "data_size": 63488 00:13:15.206 }, 00:13:15.206 { 00:13:15.206 "name": "BaseBdev3", 00:13:15.206 "uuid": "36988339-b583-5c5c-974f-fd338828ec4f", 00:13:15.206 "is_configured": true, 00:13:15.206 "data_offset": 2048, 00:13:15.206 "data_size": 63488 00:13:15.206 }, 00:13:15.206 { 00:13:15.206 "name": "BaseBdev4", 00:13:15.206 "uuid": "57e9c1d7-cb24-59ed-9066-0d5f0ec96dab", 00:13:15.206 "is_configured": true, 00:13:15.206 "data_offset": 2048, 00:13:15.206 "data_size": 63488 00:13:15.206 } 00:13:15.206 ] 00:13:15.206 }' 00:13:15.206 15:17:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:15.206 15:17:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:15.206 15:17:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:15.466 15:17:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:15.466 15:17:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:15.466 15:17:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:15.466 15:17:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:15.466 [2024-11-27 15:17:43.369016] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:15.466 15:17:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:15.466 15:17:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:13:15.466 [2024-11-27 15:17:43.439861] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:13:15.466 [2024-11-27 15:17:43.441954] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:15.466 [2024-11-27 15:17:43.564245] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:15.466 [2024-11-27 15:17:43.564795] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:15.726 [2024-11-27 15:17:43.772698] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:15.726 [2024-11-27 15:17:43.773300] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:16.296 [2024-11-27 15:17:44.112061] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:13:16.296 [2024-11-27 15:17:44.113192] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:13:16.296 164.67 IOPS, 494.00 MiB/s [2024-11-27T15:17:44.403Z] [2024-11-27 15:17:44.329079] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:16.296 [2024-11-27 15:17:44.329396] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:16.555 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:16.555 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:16.555 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:16.555 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:16.555 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:16.555 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:16.555 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:16.555 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:16.555 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:16.555 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:16.556 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:16.556 "name": "raid_bdev1", 00:13:16.556 "uuid": "0e636b34-1fee-4bd2-bbb1-1f06f968ec67", 00:13:16.556 "strip_size_kb": 0, 00:13:16.556 "state": "online", 00:13:16.556 "raid_level": "raid1", 00:13:16.556 "superblock": true, 00:13:16.556 "num_base_bdevs": 4, 00:13:16.556 "num_base_bdevs_discovered": 4, 00:13:16.556 "num_base_bdevs_operational": 4, 00:13:16.556 "process": { 00:13:16.556 "type": "rebuild", 00:13:16.556 "target": "spare", 00:13:16.556 "progress": { 00:13:16.556 "blocks": 10240, 00:13:16.556 "percent": 16 00:13:16.556 } 00:13:16.556 }, 00:13:16.556 "base_bdevs_list": [ 00:13:16.556 { 00:13:16.556 "name": "spare", 00:13:16.556 "uuid": "678e143c-d959-5d2f-9795-cd1e0ad776f6", 00:13:16.556 "is_configured": true, 00:13:16.556 "data_offset": 2048, 00:13:16.556 "data_size": 63488 00:13:16.556 }, 00:13:16.556 { 00:13:16.556 "name": "BaseBdev2", 00:13:16.556 "uuid": "3d99c079-b40c-5fd1-9da2-35620d3e4cee", 00:13:16.556 "is_configured": true, 00:13:16.556 "data_offset": 2048, 00:13:16.556 "data_size": 63488 00:13:16.556 }, 00:13:16.556 { 00:13:16.556 "name": "BaseBdev3", 00:13:16.556 "uuid": "36988339-b583-5c5c-974f-fd338828ec4f", 00:13:16.556 "is_configured": true, 00:13:16.556 "data_offset": 2048, 00:13:16.556 "data_size": 63488 00:13:16.556 }, 00:13:16.556 { 00:13:16.556 "name": "BaseBdev4", 00:13:16.556 "uuid": "57e9c1d7-cb24-59ed-9066-0d5f0ec96dab", 00:13:16.556 "is_configured": true, 00:13:16.556 "data_offset": 2048, 00:13:16.556 "data_size": 63488 00:13:16.556 } 00:13:16.556 ] 00:13:16.556 }' 00:13:16.556 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:16.556 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:16.556 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:16.556 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:16.556 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:13:16.556 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:13:16.556 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:13:16.556 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:13:16.556 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:13:16.556 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:13:16.556 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:13:16.556 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:16.556 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:16.556 [2024-11-27 15:17:44.578192] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:16.815 [2024-11-27 15:17:44.858757] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000006080 00:13:16.815 [2024-11-27 15:17:44.858866] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000006220 00:13:16.816 [2024-11-27 15:17:44.861631] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:13:16.816 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:16.816 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:13:16.816 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:13:16.816 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:16.816 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:16.816 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:16.816 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:16.816 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:16.816 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:16.816 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:16.816 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:16.816 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:16.816 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:17.075 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:17.075 "name": "raid_bdev1", 00:13:17.075 "uuid": "0e636b34-1fee-4bd2-bbb1-1f06f968ec67", 00:13:17.075 "strip_size_kb": 0, 00:13:17.075 "state": "online", 00:13:17.075 "raid_level": "raid1", 00:13:17.075 "superblock": true, 00:13:17.075 "num_base_bdevs": 4, 00:13:17.075 "num_base_bdevs_discovered": 3, 00:13:17.075 "num_base_bdevs_operational": 3, 00:13:17.075 "process": { 00:13:17.075 "type": "rebuild", 00:13:17.075 "target": "spare", 00:13:17.075 "progress": { 00:13:17.075 "blocks": 14336, 00:13:17.075 "percent": 22 00:13:17.075 } 00:13:17.075 }, 00:13:17.075 "base_bdevs_list": [ 00:13:17.075 { 00:13:17.075 "name": "spare", 00:13:17.075 "uuid": "678e143c-d959-5d2f-9795-cd1e0ad776f6", 00:13:17.075 "is_configured": true, 00:13:17.075 "data_offset": 2048, 00:13:17.075 "data_size": 63488 00:13:17.075 }, 00:13:17.075 { 00:13:17.075 "name": null, 00:13:17.075 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:17.075 "is_configured": false, 00:13:17.075 "data_offset": 0, 00:13:17.075 "data_size": 63488 00:13:17.075 }, 00:13:17.075 { 00:13:17.075 "name": "BaseBdev3", 00:13:17.075 "uuid": "36988339-b583-5c5c-974f-fd338828ec4f", 00:13:17.075 "is_configured": true, 00:13:17.075 "data_offset": 2048, 00:13:17.075 "data_size": 63488 00:13:17.075 }, 00:13:17.075 { 00:13:17.075 "name": "BaseBdev4", 00:13:17.075 "uuid": "57e9c1d7-cb24-59ed-9066-0d5f0ec96dab", 00:13:17.075 "is_configured": true, 00:13:17.075 "data_offset": 2048, 00:13:17.075 "data_size": 63488 00:13:17.075 } 00:13:17.075 ] 00:13:17.075 }' 00:13:17.075 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:17.075 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:17.075 [2024-11-27 15:17:44.982168] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:13:17.075 15:17:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:17.076 15:17:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:17.076 15:17:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@706 -- # local timeout=406 00:13:17.076 15:17:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:17.076 15:17:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:17.076 15:17:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:17.076 15:17:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:17.076 15:17:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:17.076 15:17:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:17.076 15:17:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:17.076 15:17:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:17.076 15:17:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:17.076 15:17:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:17.076 15:17:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:17.076 15:17:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:17.076 "name": "raid_bdev1", 00:13:17.076 "uuid": "0e636b34-1fee-4bd2-bbb1-1f06f968ec67", 00:13:17.076 "strip_size_kb": 0, 00:13:17.076 "state": "online", 00:13:17.076 "raid_level": "raid1", 00:13:17.076 "superblock": true, 00:13:17.076 "num_base_bdevs": 4, 00:13:17.076 "num_base_bdevs_discovered": 3, 00:13:17.076 "num_base_bdevs_operational": 3, 00:13:17.076 "process": { 00:13:17.076 "type": "rebuild", 00:13:17.076 "target": "spare", 00:13:17.076 "progress": { 00:13:17.076 "blocks": 16384, 00:13:17.076 "percent": 25 00:13:17.076 } 00:13:17.076 }, 00:13:17.076 "base_bdevs_list": [ 00:13:17.076 { 00:13:17.076 "name": "spare", 00:13:17.076 "uuid": "678e143c-d959-5d2f-9795-cd1e0ad776f6", 00:13:17.076 "is_configured": true, 00:13:17.076 "data_offset": 2048, 00:13:17.076 "data_size": 63488 00:13:17.076 }, 00:13:17.076 { 00:13:17.076 "name": null, 00:13:17.076 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:17.076 "is_configured": false, 00:13:17.076 "data_offset": 0, 00:13:17.076 "data_size": 63488 00:13:17.076 }, 00:13:17.076 { 00:13:17.076 "name": "BaseBdev3", 00:13:17.076 "uuid": "36988339-b583-5c5c-974f-fd338828ec4f", 00:13:17.076 "is_configured": true, 00:13:17.076 "data_offset": 2048, 00:13:17.076 "data_size": 63488 00:13:17.076 }, 00:13:17.076 { 00:13:17.076 "name": "BaseBdev4", 00:13:17.076 "uuid": "57e9c1d7-cb24-59ed-9066-0d5f0ec96dab", 00:13:17.076 "is_configured": true, 00:13:17.076 "data_offset": 2048, 00:13:17.076 "data_size": 63488 00:13:17.076 } 00:13:17.076 ] 00:13:17.076 }' 00:13:17.076 15:17:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:17.076 15:17:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:17.076 15:17:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:17.076 141.25 IOPS, 423.75 MiB/s [2024-11-27T15:17:45.183Z] 15:17:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:17.076 15:17:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:17.336 [2024-11-27 15:17:45.319448] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:13:18.275 125.20 IOPS, 375.60 MiB/s [2024-11-27T15:17:46.382Z] 15:17:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:18.275 15:17:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:18.275 15:17:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:18.275 15:17:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:18.275 15:17:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:18.275 15:17:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:18.275 15:17:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:18.275 15:17:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:18.275 15:17:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:18.275 15:17:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:18.275 15:17:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:18.275 15:17:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:18.275 "name": "raid_bdev1", 00:13:18.275 "uuid": "0e636b34-1fee-4bd2-bbb1-1f06f968ec67", 00:13:18.275 "strip_size_kb": 0, 00:13:18.275 "state": "online", 00:13:18.275 "raid_level": "raid1", 00:13:18.275 "superblock": true, 00:13:18.275 "num_base_bdevs": 4, 00:13:18.275 "num_base_bdevs_discovered": 3, 00:13:18.275 "num_base_bdevs_operational": 3, 00:13:18.275 "process": { 00:13:18.275 "type": "rebuild", 00:13:18.275 "target": "spare", 00:13:18.275 "progress": { 00:13:18.275 "blocks": 32768, 00:13:18.275 "percent": 51 00:13:18.275 } 00:13:18.275 }, 00:13:18.275 "base_bdevs_list": [ 00:13:18.275 { 00:13:18.275 "name": "spare", 00:13:18.275 "uuid": "678e143c-d959-5d2f-9795-cd1e0ad776f6", 00:13:18.275 "is_configured": true, 00:13:18.275 "data_offset": 2048, 00:13:18.275 "data_size": 63488 00:13:18.275 }, 00:13:18.275 { 00:13:18.275 "name": null, 00:13:18.275 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:18.275 "is_configured": false, 00:13:18.275 "data_offset": 0, 00:13:18.275 "data_size": 63488 00:13:18.275 }, 00:13:18.275 { 00:13:18.275 "name": "BaseBdev3", 00:13:18.275 "uuid": "36988339-b583-5c5c-974f-fd338828ec4f", 00:13:18.275 "is_configured": true, 00:13:18.275 "data_offset": 2048, 00:13:18.275 "data_size": 63488 00:13:18.275 }, 00:13:18.275 { 00:13:18.275 "name": "BaseBdev4", 00:13:18.275 "uuid": "57e9c1d7-cb24-59ed-9066-0d5f0ec96dab", 00:13:18.275 "is_configured": true, 00:13:18.275 "data_offset": 2048, 00:13:18.275 "data_size": 63488 00:13:18.275 } 00:13:18.275 ] 00:13:18.275 }' 00:13:18.275 15:17:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:18.275 [2024-11-27 15:17:46.260834] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:13:18.275 15:17:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:18.275 15:17:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:18.275 15:17:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:18.275 15:17:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:18.535 [2024-11-27 15:17:46.604530] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 38912 offset_begin: 36864 offset_end: 43008 00:13:19.104 [2024-11-27 15:17:47.045668] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 47104 offset_begin: 43008 offset_end: 49152 00:13:19.365 111.50 IOPS, 334.50 MiB/s [2024-11-27T15:17:47.472Z] 15:17:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:19.365 15:17:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:19.365 15:17:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:19.365 15:17:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:19.365 15:17:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:19.365 15:17:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:19.365 15:17:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:19.365 15:17:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:19.365 15:17:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:19.365 15:17:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:19.365 15:17:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:19.365 15:17:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:19.365 "name": "raid_bdev1", 00:13:19.365 "uuid": "0e636b34-1fee-4bd2-bbb1-1f06f968ec67", 00:13:19.365 "strip_size_kb": 0, 00:13:19.366 "state": "online", 00:13:19.366 "raid_level": "raid1", 00:13:19.366 "superblock": true, 00:13:19.366 "num_base_bdevs": 4, 00:13:19.366 "num_base_bdevs_discovered": 3, 00:13:19.366 "num_base_bdevs_operational": 3, 00:13:19.366 "process": { 00:13:19.366 "type": "rebuild", 00:13:19.366 "target": "spare", 00:13:19.366 "progress": { 00:13:19.366 "blocks": 49152, 00:13:19.366 "percent": 77 00:13:19.366 } 00:13:19.366 }, 00:13:19.366 "base_bdevs_list": [ 00:13:19.366 { 00:13:19.366 "name": "spare", 00:13:19.366 "uuid": "678e143c-d959-5d2f-9795-cd1e0ad776f6", 00:13:19.366 "is_configured": true, 00:13:19.366 "data_offset": 2048, 00:13:19.366 "data_size": 63488 00:13:19.366 }, 00:13:19.366 { 00:13:19.366 "name": null, 00:13:19.366 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:19.366 "is_configured": false, 00:13:19.366 "data_offset": 0, 00:13:19.366 "data_size": 63488 00:13:19.366 }, 00:13:19.366 { 00:13:19.366 "name": "BaseBdev3", 00:13:19.366 "uuid": "36988339-b583-5c5c-974f-fd338828ec4f", 00:13:19.366 "is_configured": true, 00:13:19.366 "data_offset": 2048, 00:13:19.366 "data_size": 63488 00:13:19.366 }, 00:13:19.366 { 00:13:19.366 "name": "BaseBdev4", 00:13:19.366 "uuid": "57e9c1d7-cb24-59ed-9066-0d5f0ec96dab", 00:13:19.366 "is_configured": true, 00:13:19.366 "data_offset": 2048, 00:13:19.366 "data_size": 63488 00:13:19.366 } 00:13:19.366 ] 00:13:19.366 }' 00:13:19.366 15:17:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:19.366 15:17:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:19.366 15:17:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:19.366 15:17:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:19.366 15:17:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:19.637 [2024-11-27 15:17:47.492653] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 53248 offset_begin: 49152 offset_end: 55296 00:13:20.220 101.29 IOPS, 303.86 MiB/s [2024-11-27T15:17:48.327Z] [2024-11-27 15:17:48.142735] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:13:20.220 [2024-11-27 15:17:48.247759] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:13:20.220 [2024-11-27 15:17:48.250680] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:20.480 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:20.480 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:20.480 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:20.480 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:20.480 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:20.480 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:20.480 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:20.480 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:20.480 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:20.480 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:20.480 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:20.480 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:20.480 "name": "raid_bdev1", 00:13:20.480 "uuid": "0e636b34-1fee-4bd2-bbb1-1f06f968ec67", 00:13:20.480 "strip_size_kb": 0, 00:13:20.480 "state": "online", 00:13:20.480 "raid_level": "raid1", 00:13:20.480 "superblock": true, 00:13:20.480 "num_base_bdevs": 4, 00:13:20.480 "num_base_bdevs_discovered": 3, 00:13:20.480 "num_base_bdevs_operational": 3, 00:13:20.480 "base_bdevs_list": [ 00:13:20.480 { 00:13:20.480 "name": "spare", 00:13:20.480 "uuid": "678e143c-d959-5d2f-9795-cd1e0ad776f6", 00:13:20.480 "is_configured": true, 00:13:20.480 "data_offset": 2048, 00:13:20.480 "data_size": 63488 00:13:20.480 }, 00:13:20.480 { 00:13:20.480 "name": null, 00:13:20.480 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:20.480 "is_configured": false, 00:13:20.480 "data_offset": 0, 00:13:20.480 "data_size": 63488 00:13:20.480 }, 00:13:20.480 { 00:13:20.480 "name": "BaseBdev3", 00:13:20.480 "uuid": "36988339-b583-5c5c-974f-fd338828ec4f", 00:13:20.480 "is_configured": true, 00:13:20.480 "data_offset": 2048, 00:13:20.480 "data_size": 63488 00:13:20.480 }, 00:13:20.480 { 00:13:20.480 "name": "BaseBdev4", 00:13:20.480 "uuid": "57e9c1d7-cb24-59ed-9066-0d5f0ec96dab", 00:13:20.480 "is_configured": true, 00:13:20.480 "data_offset": 2048, 00:13:20.480 "data_size": 63488 00:13:20.480 } 00:13:20.480 ] 00:13:20.480 }' 00:13:20.480 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:20.480 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:13:20.480 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@709 -- # break 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:20.740 "name": "raid_bdev1", 00:13:20.740 "uuid": "0e636b34-1fee-4bd2-bbb1-1f06f968ec67", 00:13:20.740 "strip_size_kb": 0, 00:13:20.740 "state": "online", 00:13:20.740 "raid_level": "raid1", 00:13:20.740 "superblock": true, 00:13:20.740 "num_base_bdevs": 4, 00:13:20.740 "num_base_bdevs_discovered": 3, 00:13:20.740 "num_base_bdevs_operational": 3, 00:13:20.740 "base_bdevs_list": [ 00:13:20.740 { 00:13:20.740 "name": "spare", 00:13:20.740 "uuid": "678e143c-d959-5d2f-9795-cd1e0ad776f6", 00:13:20.740 "is_configured": true, 00:13:20.740 "data_offset": 2048, 00:13:20.740 "data_size": 63488 00:13:20.740 }, 00:13:20.740 { 00:13:20.740 "name": null, 00:13:20.740 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:20.740 "is_configured": false, 00:13:20.740 "data_offset": 0, 00:13:20.740 "data_size": 63488 00:13:20.740 }, 00:13:20.740 { 00:13:20.740 "name": "BaseBdev3", 00:13:20.740 "uuid": "36988339-b583-5c5c-974f-fd338828ec4f", 00:13:20.740 "is_configured": true, 00:13:20.740 "data_offset": 2048, 00:13:20.740 "data_size": 63488 00:13:20.740 }, 00:13:20.740 { 00:13:20.740 "name": "BaseBdev4", 00:13:20.740 "uuid": "57e9c1d7-cb24-59ed-9066-0d5f0ec96dab", 00:13:20.740 "is_configured": true, 00:13:20.740 "data_offset": 2048, 00:13:20.740 "data_size": 63488 00:13:20.740 } 00:13:20.740 ] 00:13:20.740 }' 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:20.740 "name": "raid_bdev1", 00:13:20.740 "uuid": "0e636b34-1fee-4bd2-bbb1-1f06f968ec67", 00:13:20.740 "strip_size_kb": 0, 00:13:20.740 "state": "online", 00:13:20.740 "raid_level": "raid1", 00:13:20.740 "superblock": true, 00:13:20.740 "num_base_bdevs": 4, 00:13:20.740 "num_base_bdevs_discovered": 3, 00:13:20.740 "num_base_bdevs_operational": 3, 00:13:20.740 "base_bdevs_list": [ 00:13:20.740 { 00:13:20.740 "name": "spare", 00:13:20.740 "uuid": "678e143c-d959-5d2f-9795-cd1e0ad776f6", 00:13:20.740 "is_configured": true, 00:13:20.740 "data_offset": 2048, 00:13:20.740 "data_size": 63488 00:13:20.740 }, 00:13:20.740 { 00:13:20.740 "name": null, 00:13:20.740 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:20.740 "is_configured": false, 00:13:20.740 "data_offset": 0, 00:13:20.740 "data_size": 63488 00:13:20.740 }, 00:13:20.740 { 00:13:20.740 "name": "BaseBdev3", 00:13:20.740 "uuid": "36988339-b583-5c5c-974f-fd338828ec4f", 00:13:20.740 "is_configured": true, 00:13:20.740 "data_offset": 2048, 00:13:20.740 "data_size": 63488 00:13:20.740 }, 00:13:20.740 { 00:13:20.740 "name": "BaseBdev4", 00:13:20.740 "uuid": "57e9c1d7-cb24-59ed-9066-0d5f0ec96dab", 00:13:20.740 "is_configured": true, 00:13:20.740 "data_offset": 2048, 00:13:20.740 "data_size": 63488 00:13:20.740 } 00:13:20.740 ] 00:13:20.740 }' 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:20.740 15:17:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:21.308 94.12 IOPS, 282.38 MiB/s [2024-11-27T15:17:49.415Z] 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:21.308 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:21.309 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:21.309 [2024-11-27 15:17:49.202353] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:21.309 [2024-11-27 15:17:49.202392] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:21.309 00:13:21.309 Latency(us) 00:13:21.309 [2024-11-27T15:17:49.416Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:21.309 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:13:21.309 raid_bdev1 : 8.13 92.91 278.74 0.00 0.00 14788.71 270.09 115847.04 00:13:21.309 [2024-11-27T15:17:49.416Z] =================================================================================================================== 00:13:21.309 [2024-11-27T15:17:49.416Z] Total : 92.91 278.74 0.00 0.00 14788.71 270.09 115847.04 00:13:21.309 [2024-11-27 15:17:49.225293] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:21.309 [2024-11-27 15:17:49.225359] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:21.309 [2024-11-27 15:17:49.225452] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:21.309 [2024-11-27 15:17:49.225482] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:13:21.309 { 00:13:21.309 "results": [ 00:13:21.309 { 00:13:21.309 "job": "raid_bdev1", 00:13:21.309 "core_mask": "0x1", 00:13:21.309 "workload": "randrw", 00:13:21.309 "percentage": 50, 00:13:21.309 "status": "finished", 00:13:21.309 "queue_depth": 2, 00:13:21.309 "io_size": 3145728, 00:13:21.309 "runtime": 8.125989, 00:13:21.309 "iops": 92.91176741686458, 00:13:21.309 "mibps": 278.73530225059375, 00:13:21.309 "io_failed": 0, 00:13:21.309 "io_timeout": 0, 00:13:21.309 "avg_latency_us": 14788.707560079818, 00:13:21.309 "min_latency_us": 270.0855895196507, 00:13:21.309 "max_latency_us": 115847.04279475982 00:13:21.309 } 00:13:21.309 ], 00:13:21.309 "core_count": 1 00:13:21.309 } 00:13:21.309 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:21.309 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:21.309 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:21.309 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:21.309 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # jq length 00:13:21.309 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:21.309 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:13:21.309 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:13:21.309 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:13:21.309 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:13:21.309 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:21.309 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:13:21.309 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:21.309 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:13:21.309 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:21.309 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:13:21.309 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:21.309 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:21.309 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:13:21.567 /dev/nbd0 00:13:21.567 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:21.567 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:21.567 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:13:21.567 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:13:21.567 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:21.567 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:21.567 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:13:21.567 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:13:21.567 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:21.567 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:21.567 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:21.567 1+0 records in 00:13:21.567 1+0 records out 00:13:21.567 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000436763 s, 9.4 MB/s 00:13:21.567 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:21.567 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:13:21.567 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:21.567 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:21.567 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:13:21.567 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:21.567 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:21.567 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:13:21.567 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z '' ']' 00:13:21.567 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@728 -- # continue 00:13:21.567 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:13:21.567 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev3 ']' 00:13:21.567 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev3 /dev/nbd1 00:13:21.567 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:21.568 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev3') 00:13:21.568 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:21.568 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:13:21.568 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:21.568 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:13:21.568 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:21.568 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:21.568 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev3 /dev/nbd1 00:13:21.826 /dev/nbd1 00:13:21.826 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:21.826 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:21.826 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:13:21.826 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:13:21.826 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:21.826 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:21.826 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:13:21.826 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:13:21.826 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:21.826 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:21.826 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:21.826 1+0 records in 00:13:21.826 1+0 records out 00:13:21.826 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000285981 s, 14.3 MB/s 00:13:21.826 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:21.826 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:13:21.827 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:21.827 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:21.827 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:13:21.827 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:21.827 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:21.827 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:13:21.827 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:13:21.827 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:21.827 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:13:21.827 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:21.827 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:13:21.827 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:21.827 15:17:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:22.084 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:22.084 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:22.084 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:22.084 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:22.084 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:22.084 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:22.084 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:13:22.084 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:22.084 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:13:22.084 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev4 ']' 00:13:22.084 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev4 /dev/nbd1 00:13:22.084 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:22.084 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev4') 00:13:22.085 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:22.085 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:13:22.085 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:22.085 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:13:22.085 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:22.085 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:22.085 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev4 /dev/nbd1 00:13:22.344 /dev/nbd1 00:13:22.344 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:22.344 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:22.344 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:13:22.344 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:13:22.344 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:22.344 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:22.344 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:13:22.344 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:13:22.344 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:22.344 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:22.344 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:22.344 1+0 records in 00:13:22.344 1+0 records out 00:13:22.344 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000255044 s, 16.1 MB/s 00:13:22.344 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:22.344 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:13:22.344 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:22.344 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:22.344 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:13:22.344 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:22.344 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:22.344 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:13:22.344 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:13:22.344 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:22.344 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:13:22.344 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:22.344 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:13:22.344 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:22.344 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:22.602 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:22.602 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:22.602 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:22.602 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:22.602 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:22.602 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:22.602 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:13:22.602 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:22.602 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:13:22.602 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:22.602 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:13:22.602 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:22.602 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:13:22.602 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:22.602 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:22.861 [2024-11-27 15:17:50.764638] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:22.861 [2024-11-27 15:17:50.764704] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:22.861 [2024-11-27 15:17:50.764728] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:13:22.861 [2024-11-27 15:17:50.764739] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:22.861 [2024-11-27 15:17:50.766986] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:22.861 [2024-11-27 15:17:50.767022] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:22.861 [2024-11-27 15:17:50.767108] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:13:22.861 [2024-11-27 15:17:50.767155] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:22.861 [2024-11-27 15:17:50.767280] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:22.861 [2024-11-27 15:17:50.767386] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:13:22.861 spare 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:22.861 [2024-11-27 15:17:50.867304] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006600 00:13:22.861 [2024-11-27 15:17:50.867336] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:13:22.861 [2024-11-27 15:17:50.867634] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000036fc0 00:13:22.861 [2024-11-27 15:17:50.867822] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006600 00:13:22.861 [2024-11-27 15:17:50.867840] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006600 00:13:22.861 [2024-11-27 15:17:50.867999] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:22.861 "name": "raid_bdev1", 00:13:22.861 "uuid": "0e636b34-1fee-4bd2-bbb1-1f06f968ec67", 00:13:22.861 "strip_size_kb": 0, 00:13:22.861 "state": "online", 00:13:22.861 "raid_level": "raid1", 00:13:22.861 "superblock": true, 00:13:22.861 "num_base_bdevs": 4, 00:13:22.861 "num_base_bdevs_discovered": 3, 00:13:22.861 "num_base_bdevs_operational": 3, 00:13:22.861 "base_bdevs_list": [ 00:13:22.861 { 00:13:22.861 "name": "spare", 00:13:22.861 "uuid": "678e143c-d959-5d2f-9795-cd1e0ad776f6", 00:13:22.861 "is_configured": true, 00:13:22.861 "data_offset": 2048, 00:13:22.861 "data_size": 63488 00:13:22.861 }, 00:13:22.861 { 00:13:22.861 "name": null, 00:13:22.861 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:22.861 "is_configured": false, 00:13:22.861 "data_offset": 2048, 00:13:22.861 "data_size": 63488 00:13:22.861 }, 00:13:22.861 { 00:13:22.861 "name": "BaseBdev3", 00:13:22.861 "uuid": "36988339-b583-5c5c-974f-fd338828ec4f", 00:13:22.861 "is_configured": true, 00:13:22.861 "data_offset": 2048, 00:13:22.861 "data_size": 63488 00:13:22.861 }, 00:13:22.861 { 00:13:22.861 "name": "BaseBdev4", 00:13:22.861 "uuid": "57e9c1d7-cb24-59ed-9066-0d5f0ec96dab", 00:13:22.861 "is_configured": true, 00:13:22.861 "data_offset": 2048, 00:13:22.861 "data_size": 63488 00:13:22.861 } 00:13:22.861 ] 00:13:22.861 }' 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:22.861 15:17:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:23.429 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:23.429 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:23.429 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:23.429 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:23.429 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:23.429 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:23.429 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.429 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:23.429 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:23.429 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.429 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:23.429 "name": "raid_bdev1", 00:13:23.429 "uuid": "0e636b34-1fee-4bd2-bbb1-1f06f968ec67", 00:13:23.429 "strip_size_kb": 0, 00:13:23.429 "state": "online", 00:13:23.429 "raid_level": "raid1", 00:13:23.429 "superblock": true, 00:13:23.429 "num_base_bdevs": 4, 00:13:23.429 "num_base_bdevs_discovered": 3, 00:13:23.429 "num_base_bdevs_operational": 3, 00:13:23.429 "base_bdevs_list": [ 00:13:23.429 { 00:13:23.429 "name": "spare", 00:13:23.429 "uuid": "678e143c-d959-5d2f-9795-cd1e0ad776f6", 00:13:23.429 "is_configured": true, 00:13:23.429 "data_offset": 2048, 00:13:23.429 "data_size": 63488 00:13:23.429 }, 00:13:23.429 { 00:13:23.429 "name": null, 00:13:23.429 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:23.430 "is_configured": false, 00:13:23.430 "data_offset": 2048, 00:13:23.430 "data_size": 63488 00:13:23.430 }, 00:13:23.430 { 00:13:23.430 "name": "BaseBdev3", 00:13:23.430 "uuid": "36988339-b583-5c5c-974f-fd338828ec4f", 00:13:23.430 "is_configured": true, 00:13:23.430 "data_offset": 2048, 00:13:23.430 "data_size": 63488 00:13:23.430 }, 00:13:23.430 { 00:13:23.430 "name": "BaseBdev4", 00:13:23.430 "uuid": "57e9c1d7-cb24-59ed-9066-0d5f0ec96dab", 00:13:23.430 "is_configured": true, 00:13:23.430 "data_offset": 2048, 00:13:23.430 "data_size": 63488 00:13:23.430 } 00:13:23.430 ] 00:13:23.430 }' 00:13:23.430 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:23.430 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:23.430 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:23.430 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:23.430 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:23.430 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.430 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:23.430 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:13:23.430 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.430 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:13:23.430 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:23.430 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.430 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:23.430 [2024-11-27 15:17:51.515543] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:23.430 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.430 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:23.430 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:23.430 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:23.430 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:23.430 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:23.430 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:23.430 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:23.430 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:23.430 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:23.430 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:23.430 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:23.430 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.430 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:23.430 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:23.687 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.687 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:23.687 "name": "raid_bdev1", 00:13:23.687 "uuid": "0e636b34-1fee-4bd2-bbb1-1f06f968ec67", 00:13:23.687 "strip_size_kb": 0, 00:13:23.687 "state": "online", 00:13:23.687 "raid_level": "raid1", 00:13:23.687 "superblock": true, 00:13:23.687 "num_base_bdevs": 4, 00:13:23.687 "num_base_bdevs_discovered": 2, 00:13:23.687 "num_base_bdevs_operational": 2, 00:13:23.687 "base_bdevs_list": [ 00:13:23.687 { 00:13:23.687 "name": null, 00:13:23.687 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:23.687 "is_configured": false, 00:13:23.687 "data_offset": 0, 00:13:23.687 "data_size": 63488 00:13:23.687 }, 00:13:23.687 { 00:13:23.687 "name": null, 00:13:23.687 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:23.687 "is_configured": false, 00:13:23.687 "data_offset": 2048, 00:13:23.687 "data_size": 63488 00:13:23.687 }, 00:13:23.687 { 00:13:23.687 "name": "BaseBdev3", 00:13:23.687 "uuid": "36988339-b583-5c5c-974f-fd338828ec4f", 00:13:23.687 "is_configured": true, 00:13:23.687 "data_offset": 2048, 00:13:23.687 "data_size": 63488 00:13:23.687 }, 00:13:23.687 { 00:13:23.687 "name": "BaseBdev4", 00:13:23.687 "uuid": "57e9c1d7-cb24-59ed-9066-0d5f0ec96dab", 00:13:23.687 "is_configured": true, 00:13:23.687 "data_offset": 2048, 00:13:23.687 "data_size": 63488 00:13:23.687 } 00:13:23.687 ] 00:13:23.687 }' 00:13:23.687 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:23.687 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:23.946 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:23.946 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.946 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:23.946 [2024-11-27 15:17:51.982891] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:23.946 [2024-11-27 15:17:51.983106] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:13:23.946 [2024-11-27 15:17:51.983139] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:13:23.946 [2024-11-27 15:17:51.983176] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:23.946 [2024-11-27 15:17:51.987744] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000037090 00:13:23.946 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.946 15:17:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@757 -- # sleep 1 00:13:23.946 [2024-11-27 15:17:51.989598] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:25.324 15:17:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:25.324 15:17:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:25.324 15:17:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:25.324 15:17:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:25.324 15:17:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:25.324 15:17:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:25.324 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.324 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:25.324 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:25.324 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:25.324 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:25.324 "name": "raid_bdev1", 00:13:25.324 "uuid": "0e636b34-1fee-4bd2-bbb1-1f06f968ec67", 00:13:25.324 "strip_size_kb": 0, 00:13:25.324 "state": "online", 00:13:25.324 "raid_level": "raid1", 00:13:25.324 "superblock": true, 00:13:25.324 "num_base_bdevs": 4, 00:13:25.324 "num_base_bdevs_discovered": 3, 00:13:25.324 "num_base_bdevs_operational": 3, 00:13:25.324 "process": { 00:13:25.324 "type": "rebuild", 00:13:25.324 "target": "spare", 00:13:25.324 "progress": { 00:13:25.324 "blocks": 20480, 00:13:25.324 "percent": 32 00:13:25.324 } 00:13:25.324 }, 00:13:25.324 "base_bdevs_list": [ 00:13:25.324 { 00:13:25.324 "name": "spare", 00:13:25.324 "uuid": "678e143c-d959-5d2f-9795-cd1e0ad776f6", 00:13:25.324 "is_configured": true, 00:13:25.324 "data_offset": 2048, 00:13:25.324 "data_size": 63488 00:13:25.324 }, 00:13:25.324 { 00:13:25.324 "name": null, 00:13:25.324 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:25.324 "is_configured": false, 00:13:25.324 "data_offset": 2048, 00:13:25.324 "data_size": 63488 00:13:25.324 }, 00:13:25.324 { 00:13:25.324 "name": "BaseBdev3", 00:13:25.324 "uuid": "36988339-b583-5c5c-974f-fd338828ec4f", 00:13:25.324 "is_configured": true, 00:13:25.324 "data_offset": 2048, 00:13:25.324 "data_size": 63488 00:13:25.324 }, 00:13:25.324 { 00:13:25.324 "name": "BaseBdev4", 00:13:25.324 "uuid": "57e9c1d7-cb24-59ed-9066-0d5f0ec96dab", 00:13:25.324 "is_configured": true, 00:13:25.324 "data_offset": 2048, 00:13:25.324 "data_size": 63488 00:13:25.324 } 00:13:25.324 ] 00:13:25.324 }' 00:13:25.324 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:25.324 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:25.324 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:25.324 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:25.324 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:13:25.324 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.324 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:25.324 [2024-11-27 15:17:53.149939] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:25.324 [2024-11-27 15:17:53.194261] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:25.324 [2024-11-27 15:17:53.194335] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:25.324 [2024-11-27 15:17:53.194350] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:25.324 [2024-11-27 15:17:53.194360] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:25.324 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:25.324 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:25.324 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:25.324 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:25.324 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:25.325 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:25.325 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:25.325 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:25.325 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:25.325 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:25.325 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:25.325 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:25.325 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:25.325 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.325 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:25.325 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:25.325 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:25.325 "name": "raid_bdev1", 00:13:25.325 "uuid": "0e636b34-1fee-4bd2-bbb1-1f06f968ec67", 00:13:25.325 "strip_size_kb": 0, 00:13:25.325 "state": "online", 00:13:25.325 "raid_level": "raid1", 00:13:25.325 "superblock": true, 00:13:25.325 "num_base_bdevs": 4, 00:13:25.325 "num_base_bdevs_discovered": 2, 00:13:25.325 "num_base_bdevs_operational": 2, 00:13:25.325 "base_bdevs_list": [ 00:13:25.325 { 00:13:25.325 "name": null, 00:13:25.325 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:25.325 "is_configured": false, 00:13:25.325 "data_offset": 0, 00:13:25.325 "data_size": 63488 00:13:25.325 }, 00:13:25.325 { 00:13:25.325 "name": null, 00:13:25.325 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:25.325 "is_configured": false, 00:13:25.325 "data_offset": 2048, 00:13:25.325 "data_size": 63488 00:13:25.325 }, 00:13:25.325 { 00:13:25.325 "name": "BaseBdev3", 00:13:25.325 "uuid": "36988339-b583-5c5c-974f-fd338828ec4f", 00:13:25.325 "is_configured": true, 00:13:25.325 "data_offset": 2048, 00:13:25.325 "data_size": 63488 00:13:25.325 }, 00:13:25.325 { 00:13:25.325 "name": "BaseBdev4", 00:13:25.325 "uuid": "57e9c1d7-cb24-59ed-9066-0d5f0ec96dab", 00:13:25.325 "is_configured": true, 00:13:25.325 "data_offset": 2048, 00:13:25.325 "data_size": 63488 00:13:25.325 } 00:13:25.325 ] 00:13:25.325 }' 00:13:25.325 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:25.325 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:25.584 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:25.585 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.585 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:25.585 [2024-11-27 15:17:53.650103] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:25.585 [2024-11-27 15:17:53.650181] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:25.585 [2024-11-27 15:17:53.650208] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c680 00:13:25.585 [2024-11-27 15:17:53.650220] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:25.585 [2024-11-27 15:17:53.650690] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:25.585 [2024-11-27 15:17:53.650716] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:25.585 [2024-11-27 15:17:53.650806] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:13:25.585 [2024-11-27 15:17:53.650827] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:13:25.585 [2024-11-27 15:17:53.650836] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:13:25.585 [2024-11-27 15:17:53.650860] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:25.585 [2024-11-27 15:17:53.655487] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000037160 00:13:25.585 spare 00:13:25.585 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:25.585 [2024-11-27 15:17:53.657321] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:25.585 15:17:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@764 -- # sleep 1 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:26.966 "name": "raid_bdev1", 00:13:26.966 "uuid": "0e636b34-1fee-4bd2-bbb1-1f06f968ec67", 00:13:26.966 "strip_size_kb": 0, 00:13:26.966 "state": "online", 00:13:26.966 "raid_level": "raid1", 00:13:26.966 "superblock": true, 00:13:26.966 "num_base_bdevs": 4, 00:13:26.966 "num_base_bdevs_discovered": 3, 00:13:26.966 "num_base_bdevs_operational": 3, 00:13:26.966 "process": { 00:13:26.966 "type": "rebuild", 00:13:26.966 "target": "spare", 00:13:26.966 "progress": { 00:13:26.966 "blocks": 20480, 00:13:26.966 "percent": 32 00:13:26.966 } 00:13:26.966 }, 00:13:26.966 "base_bdevs_list": [ 00:13:26.966 { 00:13:26.966 "name": "spare", 00:13:26.966 "uuid": "678e143c-d959-5d2f-9795-cd1e0ad776f6", 00:13:26.966 "is_configured": true, 00:13:26.966 "data_offset": 2048, 00:13:26.966 "data_size": 63488 00:13:26.966 }, 00:13:26.966 { 00:13:26.966 "name": null, 00:13:26.966 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:26.966 "is_configured": false, 00:13:26.966 "data_offset": 2048, 00:13:26.966 "data_size": 63488 00:13:26.966 }, 00:13:26.966 { 00:13:26.966 "name": "BaseBdev3", 00:13:26.966 "uuid": "36988339-b583-5c5c-974f-fd338828ec4f", 00:13:26.966 "is_configured": true, 00:13:26.966 "data_offset": 2048, 00:13:26.966 "data_size": 63488 00:13:26.966 }, 00:13:26.966 { 00:13:26.966 "name": "BaseBdev4", 00:13:26.966 "uuid": "57e9c1d7-cb24-59ed-9066-0d5f0ec96dab", 00:13:26.966 "is_configured": true, 00:13:26.966 "data_offset": 2048, 00:13:26.966 "data_size": 63488 00:13:26.966 } 00:13:26.966 ] 00:13:26.966 }' 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:26.966 [2024-11-27 15:17:54.821807] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:26.966 [2024-11-27 15:17:54.861682] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:26.966 [2024-11-27 15:17:54.861752] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:26.966 [2024-11-27 15:17:54.861770] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:26.966 [2024-11-27 15:17:54.861777] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:26.966 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:26.966 "name": "raid_bdev1", 00:13:26.966 "uuid": "0e636b34-1fee-4bd2-bbb1-1f06f968ec67", 00:13:26.966 "strip_size_kb": 0, 00:13:26.966 "state": "online", 00:13:26.966 "raid_level": "raid1", 00:13:26.966 "superblock": true, 00:13:26.966 "num_base_bdevs": 4, 00:13:26.966 "num_base_bdevs_discovered": 2, 00:13:26.966 "num_base_bdevs_operational": 2, 00:13:26.966 "base_bdevs_list": [ 00:13:26.966 { 00:13:26.966 "name": null, 00:13:26.967 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:26.967 "is_configured": false, 00:13:26.967 "data_offset": 0, 00:13:26.967 "data_size": 63488 00:13:26.967 }, 00:13:26.967 { 00:13:26.967 "name": null, 00:13:26.967 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:26.967 "is_configured": false, 00:13:26.967 "data_offset": 2048, 00:13:26.967 "data_size": 63488 00:13:26.967 }, 00:13:26.967 { 00:13:26.967 "name": "BaseBdev3", 00:13:26.967 "uuid": "36988339-b583-5c5c-974f-fd338828ec4f", 00:13:26.967 "is_configured": true, 00:13:26.967 "data_offset": 2048, 00:13:26.967 "data_size": 63488 00:13:26.967 }, 00:13:26.967 { 00:13:26.967 "name": "BaseBdev4", 00:13:26.967 "uuid": "57e9c1d7-cb24-59ed-9066-0d5f0ec96dab", 00:13:26.967 "is_configured": true, 00:13:26.967 "data_offset": 2048, 00:13:26.967 "data_size": 63488 00:13:26.967 } 00:13:26.967 ] 00:13:26.967 }' 00:13:26.967 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:26.967 15:17:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:27.227 15:17:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:27.227 15:17:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:27.227 15:17:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:27.227 15:17:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:27.227 15:17:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:27.227 15:17:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:27.227 15:17:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:27.227 15:17:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.227 15:17:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:27.487 15:17:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.487 15:17:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:27.487 "name": "raid_bdev1", 00:13:27.487 "uuid": "0e636b34-1fee-4bd2-bbb1-1f06f968ec67", 00:13:27.487 "strip_size_kb": 0, 00:13:27.487 "state": "online", 00:13:27.487 "raid_level": "raid1", 00:13:27.487 "superblock": true, 00:13:27.488 "num_base_bdevs": 4, 00:13:27.488 "num_base_bdevs_discovered": 2, 00:13:27.488 "num_base_bdevs_operational": 2, 00:13:27.488 "base_bdevs_list": [ 00:13:27.488 { 00:13:27.488 "name": null, 00:13:27.488 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:27.488 "is_configured": false, 00:13:27.488 "data_offset": 0, 00:13:27.488 "data_size": 63488 00:13:27.488 }, 00:13:27.488 { 00:13:27.488 "name": null, 00:13:27.488 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:27.488 "is_configured": false, 00:13:27.488 "data_offset": 2048, 00:13:27.488 "data_size": 63488 00:13:27.488 }, 00:13:27.488 { 00:13:27.488 "name": "BaseBdev3", 00:13:27.488 "uuid": "36988339-b583-5c5c-974f-fd338828ec4f", 00:13:27.488 "is_configured": true, 00:13:27.488 "data_offset": 2048, 00:13:27.488 "data_size": 63488 00:13:27.488 }, 00:13:27.488 { 00:13:27.488 "name": "BaseBdev4", 00:13:27.488 "uuid": "57e9c1d7-cb24-59ed-9066-0d5f0ec96dab", 00:13:27.488 "is_configured": true, 00:13:27.488 "data_offset": 2048, 00:13:27.488 "data_size": 63488 00:13:27.488 } 00:13:27.488 ] 00:13:27.488 }' 00:13:27.488 15:17:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:27.488 15:17:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:27.488 15:17:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:27.488 15:17:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:27.488 15:17:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:13:27.488 15:17:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.488 15:17:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:27.488 15:17:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.488 15:17:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:27.488 15:17:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.488 15:17:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:27.488 [2024-11-27 15:17:55.457441] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:27.488 [2024-11-27 15:17:55.457502] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:27.488 [2024-11-27 15:17:55.457526] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000cc80 00:13:27.488 [2024-11-27 15:17:55.457535] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:27.488 [2024-11-27 15:17:55.457990] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:27.488 [2024-11-27 15:17:55.458009] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:27.488 [2024-11-27 15:17:55.458087] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:13:27.488 [2024-11-27 15:17:55.458110] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:13:27.488 [2024-11-27 15:17:55.458122] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:13:27.488 [2024-11-27 15:17:55.458134] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:13:27.488 BaseBdev1 00:13:27.488 15:17:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.488 15:17:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@775 -- # sleep 1 00:13:28.428 15:17:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:28.428 15:17:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:28.428 15:17:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:28.428 15:17:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:28.428 15:17:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:28.428 15:17:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:28.428 15:17:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:28.428 15:17:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:28.428 15:17:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:28.428 15:17:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:28.428 15:17:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:28.428 15:17:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:28.428 15:17:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:28.428 15:17:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:28.428 15:17:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:28.428 15:17:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:28.428 "name": "raid_bdev1", 00:13:28.428 "uuid": "0e636b34-1fee-4bd2-bbb1-1f06f968ec67", 00:13:28.428 "strip_size_kb": 0, 00:13:28.428 "state": "online", 00:13:28.428 "raid_level": "raid1", 00:13:28.428 "superblock": true, 00:13:28.428 "num_base_bdevs": 4, 00:13:28.428 "num_base_bdevs_discovered": 2, 00:13:28.428 "num_base_bdevs_operational": 2, 00:13:28.428 "base_bdevs_list": [ 00:13:28.428 { 00:13:28.428 "name": null, 00:13:28.428 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:28.428 "is_configured": false, 00:13:28.428 "data_offset": 0, 00:13:28.428 "data_size": 63488 00:13:28.428 }, 00:13:28.428 { 00:13:28.428 "name": null, 00:13:28.428 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:28.428 "is_configured": false, 00:13:28.428 "data_offset": 2048, 00:13:28.428 "data_size": 63488 00:13:28.428 }, 00:13:28.428 { 00:13:28.428 "name": "BaseBdev3", 00:13:28.428 "uuid": "36988339-b583-5c5c-974f-fd338828ec4f", 00:13:28.428 "is_configured": true, 00:13:28.428 "data_offset": 2048, 00:13:28.428 "data_size": 63488 00:13:28.428 }, 00:13:28.428 { 00:13:28.428 "name": "BaseBdev4", 00:13:28.428 "uuid": "57e9c1d7-cb24-59ed-9066-0d5f0ec96dab", 00:13:28.429 "is_configured": true, 00:13:28.429 "data_offset": 2048, 00:13:28.429 "data_size": 63488 00:13:28.429 } 00:13:28.429 ] 00:13:28.429 }' 00:13:28.429 15:17:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:28.429 15:17:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:28.998 15:17:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:28.998 15:17:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:28.998 15:17:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:28.998 15:17:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:28.998 15:17:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:28.998 15:17:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:28.998 15:17:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:28.998 15:17:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:28.998 15:17:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:28.998 15:17:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:28.998 15:17:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:28.998 "name": "raid_bdev1", 00:13:28.998 "uuid": "0e636b34-1fee-4bd2-bbb1-1f06f968ec67", 00:13:28.998 "strip_size_kb": 0, 00:13:28.998 "state": "online", 00:13:28.998 "raid_level": "raid1", 00:13:28.998 "superblock": true, 00:13:28.998 "num_base_bdevs": 4, 00:13:28.998 "num_base_bdevs_discovered": 2, 00:13:28.998 "num_base_bdevs_operational": 2, 00:13:28.998 "base_bdevs_list": [ 00:13:28.998 { 00:13:28.998 "name": null, 00:13:28.998 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:28.998 "is_configured": false, 00:13:28.998 "data_offset": 0, 00:13:28.998 "data_size": 63488 00:13:28.998 }, 00:13:28.998 { 00:13:28.998 "name": null, 00:13:28.998 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:28.998 "is_configured": false, 00:13:28.998 "data_offset": 2048, 00:13:28.998 "data_size": 63488 00:13:28.998 }, 00:13:28.998 { 00:13:28.998 "name": "BaseBdev3", 00:13:28.998 "uuid": "36988339-b583-5c5c-974f-fd338828ec4f", 00:13:28.998 "is_configured": true, 00:13:28.998 "data_offset": 2048, 00:13:28.998 "data_size": 63488 00:13:28.998 }, 00:13:28.998 { 00:13:28.998 "name": "BaseBdev4", 00:13:28.998 "uuid": "57e9c1d7-cb24-59ed-9066-0d5f0ec96dab", 00:13:28.998 "is_configured": true, 00:13:28.998 "data_offset": 2048, 00:13:28.998 "data_size": 63488 00:13:28.998 } 00:13:28.998 ] 00:13:28.998 }' 00:13:28.998 15:17:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:28.998 15:17:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:28.998 15:17:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:29.258 15:17:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:29.258 15:17:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:29.258 15:17:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@652 -- # local es=0 00:13:29.258 15:17:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:29.258 15:17:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:13:29.258 15:17:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:13:29.258 15:17:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:13:29.258 15:17:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:13:29.258 15:17:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:29.258 15:17:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:29.258 15:17:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:29.258 [2024-11-27 15:17:57.123530] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:29.258 [2024-11-27 15:17:57.123729] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:13:29.258 [2024-11-27 15:17:57.123754] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:13:29.258 request: 00:13:29.258 { 00:13:29.258 "base_bdev": "BaseBdev1", 00:13:29.258 "raid_bdev": "raid_bdev1", 00:13:29.258 "method": "bdev_raid_add_base_bdev", 00:13:29.258 "req_id": 1 00:13:29.258 } 00:13:29.258 Got JSON-RPC error response 00:13:29.258 response: 00:13:29.258 { 00:13:29.258 "code": -22, 00:13:29.258 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:13:29.258 } 00:13:29.258 15:17:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:13:29.258 15:17:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@655 -- # es=1 00:13:29.258 15:17:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:13:29.258 15:17:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:13:29.258 15:17:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:13:29.258 15:17:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@779 -- # sleep 1 00:13:30.198 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:30.198 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:30.198 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:30.198 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:30.198 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:30.198 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:30.198 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:30.198 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:30.198 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:30.198 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:30.198 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:30.198 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:30.198 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:30.198 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:30.198 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:30.198 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:30.198 "name": "raid_bdev1", 00:13:30.198 "uuid": "0e636b34-1fee-4bd2-bbb1-1f06f968ec67", 00:13:30.198 "strip_size_kb": 0, 00:13:30.198 "state": "online", 00:13:30.198 "raid_level": "raid1", 00:13:30.198 "superblock": true, 00:13:30.198 "num_base_bdevs": 4, 00:13:30.198 "num_base_bdevs_discovered": 2, 00:13:30.198 "num_base_bdevs_operational": 2, 00:13:30.198 "base_bdevs_list": [ 00:13:30.198 { 00:13:30.198 "name": null, 00:13:30.198 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:30.198 "is_configured": false, 00:13:30.198 "data_offset": 0, 00:13:30.198 "data_size": 63488 00:13:30.198 }, 00:13:30.198 { 00:13:30.198 "name": null, 00:13:30.198 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:30.198 "is_configured": false, 00:13:30.198 "data_offset": 2048, 00:13:30.198 "data_size": 63488 00:13:30.198 }, 00:13:30.198 { 00:13:30.198 "name": "BaseBdev3", 00:13:30.198 "uuid": "36988339-b583-5c5c-974f-fd338828ec4f", 00:13:30.198 "is_configured": true, 00:13:30.198 "data_offset": 2048, 00:13:30.198 "data_size": 63488 00:13:30.198 }, 00:13:30.198 { 00:13:30.198 "name": "BaseBdev4", 00:13:30.198 "uuid": "57e9c1d7-cb24-59ed-9066-0d5f0ec96dab", 00:13:30.198 "is_configured": true, 00:13:30.198 "data_offset": 2048, 00:13:30.198 "data_size": 63488 00:13:30.198 } 00:13:30.198 ] 00:13:30.198 }' 00:13:30.198 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:30.198 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:30.766 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:30.766 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:30.766 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:30.766 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:30.766 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:30.766 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:30.766 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:30.766 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:30.766 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:30.766 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:30.766 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:30.766 "name": "raid_bdev1", 00:13:30.766 "uuid": "0e636b34-1fee-4bd2-bbb1-1f06f968ec67", 00:13:30.766 "strip_size_kb": 0, 00:13:30.766 "state": "online", 00:13:30.766 "raid_level": "raid1", 00:13:30.766 "superblock": true, 00:13:30.766 "num_base_bdevs": 4, 00:13:30.766 "num_base_bdevs_discovered": 2, 00:13:30.766 "num_base_bdevs_operational": 2, 00:13:30.766 "base_bdevs_list": [ 00:13:30.766 { 00:13:30.766 "name": null, 00:13:30.766 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:30.766 "is_configured": false, 00:13:30.766 "data_offset": 0, 00:13:30.766 "data_size": 63488 00:13:30.766 }, 00:13:30.766 { 00:13:30.766 "name": null, 00:13:30.766 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:30.766 "is_configured": false, 00:13:30.766 "data_offset": 2048, 00:13:30.766 "data_size": 63488 00:13:30.766 }, 00:13:30.766 { 00:13:30.766 "name": "BaseBdev3", 00:13:30.766 "uuid": "36988339-b583-5c5c-974f-fd338828ec4f", 00:13:30.766 "is_configured": true, 00:13:30.766 "data_offset": 2048, 00:13:30.766 "data_size": 63488 00:13:30.766 }, 00:13:30.766 { 00:13:30.766 "name": "BaseBdev4", 00:13:30.766 "uuid": "57e9c1d7-cb24-59ed-9066-0d5f0ec96dab", 00:13:30.766 "is_configured": true, 00:13:30.766 "data_offset": 2048, 00:13:30.766 "data_size": 63488 00:13:30.766 } 00:13:30.766 ] 00:13:30.766 }' 00:13:30.766 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:30.766 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:30.766 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:30.766 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:30.766 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@784 -- # killprocess 89929 00:13:30.766 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@954 -- # '[' -z 89929 ']' 00:13:30.766 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@958 -- # kill -0 89929 00:13:30.766 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@959 -- # uname 00:13:30.766 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:30.766 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 89929 00:13:30.766 killing process with pid 89929 00:13:30.766 Received shutdown signal, test time was about 17.687466 seconds 00:13:30.766 00:13:30.766 Latency(us) 00:13:30.766 [2024-11-27T15:17:58.873Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:30.766 [2024-11-27T15:17:58.873Z] =================================================================================================================== 00:13:30.766 [2024-11-27T15:17:58.873Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:13:30.766 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:30.766 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:30.766 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@972 -- # echo 'killing process with pid 89929' 00:13:30.766 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@973 -- # kill 89929 00:13:30.766 [2024-11-27 15:17:58.765231] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:30.766 15:17:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@978 -- # wait 89929 00:13:30.766 [2024-11-27 15:17:58.765374] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:30.766 [2024-11-27 15:17:58.765446] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:30.766 [2024-11-27 15:17:58.765464] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state offline 00:13:30.766 [2024-11-27 15:17:58.812771] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:31.027 15:17:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@786 -- # return 0 00:13:31.027 00:13:31.027 real 0m19.718s 00:13:31.027 user 0m26.390s 00:13:31.027 sys 0m2.563s 00:13:31.027 15:17:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:31.027 15:17:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:31.027 ************************************ 00:13:31.027 END TEST raid_rebuild_test_sb_io 00:13:31.027 ************************************ 00:13:31.027 15:17:59 bdev_raid -- bdev/bdev_raid.sh@985 -- # for n in {3..4} 00:13:31.027 15:17:59 bdev_raid -- bdev/bdev_raid.sh@986 -- # run_test raid5f_state_function_test raid_state_function_test raid5f 3 false 00:13:31.027 15:17:59 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:13:31.027 15:17:59 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:31.027 15:17:59 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:31.027 ************************************ 00:13:31.027 START TEST raid5f_state_function_test 00:13:31.027 ************************************ 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid5f 3 false 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=90634 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 90634' 00:13:31.027 Process raid pid: 90634 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 90634 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 90634 ']' 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:31.027 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:13:31.027 15:17:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:31.287 [2024-11-27 15:17:59.193625] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:13:31.287 [2024-11-27 15:17:59.193852] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:31.287 [2024-11-27 15:17:59.362604] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:31.287 [2024-11-27 15:17:59.391391] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:31.546 [2024-11-27 15:17:59.436450] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:31.546 [2024-11-27 15:17:59.436583] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:32.116 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:13:32.116 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:13:32.116 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:32.116 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.116 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.116 [2024-11-27 15:18:00.024704] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:32.116 [2024-11-27 15:18:00.024827] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:32.116 [2024-11-27 15:18:00.024868] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:32.116 [2024-11-27 15:18:00.024910] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:32.116 [2024-11-27 15:18:00.024950] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:32.116 [2024-11-27 15:18:00.024979] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:32.116 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.116 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:32.116 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:32.116 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:32.116 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:32.116 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:32.116 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:32.116 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:32.116 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:32.116 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:32.116 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:32.116 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:32.116 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.116 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.116 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:32.116 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.116 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:32.116 "name": "Existed_Raid", 00:13:32.116 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:32.116 "strip_size_kb": 64, 00:13:32.116 "state": "configuring", 00:13:32.116 "raid_level": "raid5f", 00:13:32.116 "superblock": false, 00:13:32.116 "num_base_bdevs": 3, 00:13:32.116 "num_base_bdevs_discovered": 0, 00:13:32.117 "num_base_bdevs_operational": 3, 00:13:32.117 "base_bdevs_list": [ 00:13:32.117 { 00:13:32.117 "name": "BaseBdev1", 00:13:32.117 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:32.117 "is_configured": false, 00:13:32.117 "data_offset": 0, 00:13:32.117 "data_size": 0 00:13:32.117 }, 00:13:32.117 { 00:13:32.117 "name": "BaseBdev2", 00:13:32.117 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:32.117 "is_configured": false, 00:13:32.117 "data_offset": 0, 00:13:32.117 "data_size": 0 00:13:32.117 }, 00:13:32.117 { 00:13:32.117 "name": "BaseBdev3", 00:13:32.117 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:32.117 "is_configured": false, 00:13:32.117 "data_offset": 0, 00:13:32.117 "data_size": 0 00:13:32.117 } 00:13:32.117 ] 00:13:32.117 }' 00:13:32.117 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:32.117 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.708 [2024-11-27 15:18:00.515763] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:32.708 [2024-11-27 15:18:00.515852] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.708 [2024-11-27 15:18:00.527748] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:32.708 [2024-11-27 15:18:00.527792] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:32.708 [2024-11-27 15:18:00.527802] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:32.708 [2024-11-27 15:18:00.527812] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:32.708 [2024-11-27 15:18:00.527818] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:32.708 [2024-11-27 15:18:00.527826] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.708 [2024-11-27 15:18:00.548778] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:32.708 BaseBdev1 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.708 [ 00:13:32.708 { 00:13:32.708 "name": "BaseBdev1", 00:13:32.708 "aliases": [ 00:13:32.708 "fdf93e0d-63c3-4037-a351-e9185cc8eb77" 00:13:32.708 ], 00:13:32.708 "product_name": "Malloc disk", 00:13:32.708 "block_size": 512, 00:13:32.708 "num_blocks": 65536, 00:13:32.708 "uuid": "fdf93e0d-63c3-4037-a351-e9185cc8eb77", 00:13:32.708 "assigned_rate_limits": { 00:13:32.708 "rw_ios_per_sec": 0, 00:13:32.708 "rw_mbytes_per_sec": 0, 00:13:32.708 "r_mbytes_per_sec": 0, 00:13:32.708 "w_mbytes_per_sec": 0 00:13:32.708 }, 00:13:32.708 "claimed": true, 00:13:32.708 "claim_type": "exclusive_write", 00:13:32.708 "zoned": false, 00:13:32.708 "supported_io_types": { 00:13:32.708 "read": true, 00:13:32.708 "write": true, 00:13:32.708 "unmap": true, 00:13:32.708 "flush": true, 00:13:32.708 "reset": true, 00:13:32.708 "nvme_admin": false, 00:13:32.708 "nvme_io": false, 00:13:32.708 "nvme_io_md": false, 00:13:32.708 "write_zeroes": true, 00:13:32.708 "zcopy": true, 00:13:32.708 "get_zone_info": false, 00:13:32.708 "zone_management": false, 00:13:32.708 "zone_append": false, 00:13:32.708 "compare": false, 00:13:32.708 "compare_and_write": false, 00:13:32.708 "abort": true, 00:13:32.708 "seek_hole": false, 00:13:32.708 "seek_data": false, 00:13:32.708 "copy": true, 00:13:32.708 "nvme_iov_md": false 00:13:32.708 }, 00:13:32.708 "memory_domains": [ 00:13:32.708 { 00:13:32.708 "dma_device_id": "system", 00:13:32.708 "dma_device_type": 1 00:13:32.708 }, 00:13:32.708 { 00:13:32.708 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:32.708 "dma_device_type": 2 00:13:32.708 } 00:13:32.708 ], 00:13:32.708 "driver_specific": {} 00:13:32.708 } 00:13:32.708 ] 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.708 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:32.708 "name": "Existed_Raid", 00:13:32.708 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:32.708 "strip_size_kb": 64, 00:13:32.708 "state": "configuring", 00:13:32.708 "raid_level": "raid5f", 00:13:32.708 "superblock": false, 00:13:32.708 "num_base_bdevs": 3, 00:13:32.708 "num_base_bdevs_discovered": 1, 00:13:32.708 "num_base_bdevs_operational": 3, 00:13:32.708 "base_bdevs_list": [ 00:13:32.708 { 00:13:32.708 "name": "BaseBdev1", 00:13:32.708 "uuid": "fdf93e0d-63c3-4037-a351-e9185cc8eb77", 00:13:32.708 "is_configured": true, 00:13:32.708 "data_offset": 0, 00:13:32.708 "data_size": 65536 00:13:32.708 }, 00:13:32.708 { 00:13:32.708 "name": "BaseBdev2", 00:13:32.708 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:32.708 "is_configured": false, 00:13:32.708 "data_offset": 0, 00:13:32.708 "data_size": 0 00:13:32.708 }, 00:13:32.709 { 00:13:32.709 "name": "BaseBdev3", 00:13:32.709 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:32.709 "is_configured": false, 00:13:32.709 "data_offset": 0, 00:13:32.709 "data_size": 0 00:13:32.709 } 00:13:32.709 ] 00:13:32.709 }' 00:13:32.709 15:18:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:32.709 15:18:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.968 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:32.968 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.968 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.968 [2024-11-27 15:18:01.039984] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:32.968 [2024-11-27 15:18:01.040084] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:13:32.968 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.968 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:32.968 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.968 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.968 [2024-11-27 15:18:01.052020] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:32.968 [2024-11-27 15:18:01.053977] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:32.968 [2024-11-27 15:18:01.054069] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:32.968 [2024-11-27 15:18:01.054099] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:32.968 [2024-11-27 15:18:01.054124] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:32.968 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.968 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:13:32.968 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:32.968 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:32.968 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:32.968 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:32.968 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:32.968 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:32.968 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:32.968 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:32.968 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:32.968 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:32.968 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:32.968 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:32.968 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:32.968 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.968 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:33.227 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:33.228 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:33.228 "name": "Existed_Raid", 00:13:33.228 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:33.228 "strip_size_kb": 64, 00:13:33.228 "state": "configuring", 00:13:33.228 "raid_level": "raid5f", 00:13:33.228 "superblock": false, 00:13:33.228 "num_base_bdevs": 3, 00:13:33.228 "num_base_bdevs_discovered": 1, 00:13:33.228 "num_base_bdevs_operational": 3, 00:13:33.228 "base_bdevs_list": [ 00:13:33.228 { 00:13:33.228 "name": "BaseBdev1", 00:13:33.228 "uuid": "fdf93e0d-63c3-4037-a351-e9185cc8eb77", 00:13:33.228 "is_configured": true, 00:13:33.228 "data_offset": 0, 00:13:33.228 "data_size": 65536 00:13:33.228 }, 00:13:33.228 { 00:13:33.228 "name": "BaseBdev2", 00:13:33.228 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:33.228 "is_configured": false, 00:13:33.228 "data_offset": 0, 00:13:33.228 "data_size": 0 00:13:33.228 }, 00:13:33.228 { 00:13:33.228 "name": "BaseBdev3", 00:13:33.228 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:33.228 "is_configured": false, 00:13:33.228 "data_offset": 0, 00:13:33.228 "data_size": 0 00:13:33.228 } 00:13:33.228 ] 00:13:33.228 }' 00:13:33.228 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:33.228 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:33.488 [2024-11-27 15:18:01.498579] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:33.488 BaseBdev2 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:33.488 [ 00:13:33.488 { 00:13:33.488 "name": "BaseBdev2", 00:13:33.488 "aliases": [ 00:13:33.488 "49fee062-8804-44ee-864b-ccf62b791e0a" 00:13:33.488 ], 00:13:33.488 "product_name": "Malloc disk", 00:13:33.488 "block_size": 512, 00:13:33.488 "num_blocks": 65536, 00:13:33.488 "uuid": "49fee062-8804-44ee-864b-ccf62b791e0a", 00:13:33.488 "assigned_rate_limits": { 00:13:33.488 "rw_ios_per_sec": 0, 00:13:33.488 "rw_mbytes_per_sec": 0, 00:13:33.488 "r_mbytes_per_sec": 0, 00:13:33.488 "w_mbytes_per_sec": 0 00:13:33.488 }, 00:13:33.488 "claimed": true, 00:13:33.488 "claim_type": "exclusive_write", 00:13:33.488 "zoned": false, 00:13:33.488 "supported_io_types": { 00:13:33.488 "read": true, 00:13:33.488 "write": true, 00:13:33.488 "unmap": true, 00:13:33.488 "flush": true, 00:13:33.488 "reset": true, 00:13:33.488 "nvme_admin": false, 00:13:33.488 "nvme_io": false, 00:13:33.488 "nvme_io_md": false, 00:13:33.488 "write_zeroes": true, 00:13:33.488 "zcopy": true, 00:13:33.488 "get_zone_info": false, 00:13:33.488 "zone_management": false, 00:13:33.488 "zone_append": false, 00:13:33.488 "compare": false, 00:13:33.488 "compare_and_write": false, 00:13:33.488 "abort": true, 00:13:33.488 "seek_hole": false, 00:13:33.488 "seek_data": false, 00:13:33.488 "copy": true, 00:13:33.488 "nvme_iov_md": false 00:13:33.488 }, 00:13:33.488 "memory_domains": [ 00:13:33.488 { 00:13:33.488 "dma_device_id": "system", 00:13:33.488 "dma_device_type": 1 00:13:33.488 }, 00:13:33.488 { 00:13:33.488 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:33.488 "dma_device_type": 2 00:13:33.488 } 00:13:33.488 ], 00:13:33.488 "driver_specific": {} 00:13:33.488 } 00:13:33.488 ] 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:33.488 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:33.747 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:33.747 "name": "Existed_Raid", 00:13:33.747 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:33.747 "strip_size_kb": 64, 00:13:33.747 "state": "configuring", 00:13:33.747 "raid_level": "raid5f", 00:13:33.747 "superblock": false, 00:13:33.747 "num_base_bdevs": 3, 00:13:33.747 "num_base_bdevs_discovered": 2, 00:13:33.747 "num_base_bdevs_operational": 3, 00:13:33.747 "base_bdevs_list": [ 00:13:33.747 { 00:13:33.747 "name": "BaseBdev1", 00:13:33.747 "uuid": "fdf93e0d-63c3-4037-a351-e9185cc8eb77", 00:13:33.747 "is_configured": true, 00:13:33.747 "data_offset": 0, 00:13:33.747 "data_size": 65536 00:13:33.747 }, 00:13:33.747 { 00:13:33.747 "name": "BaseBdev2", 00:13:33.747 "uuid": "49fee062-8804-44ee-864b-ccf62b791e0a", 00:13:33.747 "is_configured": true, 00:13:33.747 "data_offset": 0, 00:13:33.747 "data_size": 65536 00:13:33.747 }, 00:13:33.747 { 00:13:33.747 "name": "BaseBdev3", 00:13:33.747 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:33.747 "is_configured": false, 00:13:33.747 "data_offset": 0, 00:13:33.747 "data_size": 0 00:13:33.747 } 00:13:33.747 ] 00:13:33.747 }' 00:13:33.747 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:33.747 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:34.007 15:18:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:13:34.007 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.007 15:18:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:34.007 [2024-11-27 15:18:02.002951] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:34.007 [2024-11-27 15:18:02.003124] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:13:34.007 [2024-11-27 15:18:02.003178] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:13:34.007 [2024-11-27 15:18:02.003580] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:13:34.007 [2024-11-27 15:18:02.004206] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:13:34.007 [2024-11-27 15:18:02.004269] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:13:34.007 [2024-11-27 15:18:02.004558] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:34.007 BaseBdev3 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:34.007 [ 00:13:34.007 { 00:13:34.007 "name": "BaseBdev3", 00:13:34.007 "aliases": [ 00:13:34.007 "4815ca72-eec5-419e-9508-41cc45f5e7dc" 00:13:34.007 ], 00:13:34.007 "product_name": "Malloc disk", 00:13:34.007 "block_size": 512, 00:13:34.007 "num_blocks": 65536, 00:13:34.007 "uuid": "4815ca72-eec5-419e-9508-41cc45f5e7dc", 00:13:34.007 "assigned_rate_limits": { 00:13:34.007 "rw_ios_per_sec": 0, 00:13:34.007 "rw_mbytes_per_sec": 0, 00:13:34.007 "r_mbytes_per_sec": 0, 00:13:34.007 "w_mbytes_per_sec": 0 00:13:34.007 }, 00:13:34.007 "claimed": true, 00:13:34.007 "claim_type": "exclusive_write", 00:13:34.007 "zoned": false, 00:13:34.007 "supported_io_types": { 00:13:34.007 "read": true, 00:13:34.007 "write": true, 00:13:34.007 "unmap": true, 00:13:34.007 "flush": true, 00:13:34.007 "reset": true, 00:13:34.007 "nvme_admin": false, 00:13:34.007 "nvme_io": false, 00:13:34.007 "nvme_io_md": false, 00:13:34.007 "write_zeroes": true, 00:13:34.007 "zcopy": true, 00:13:34.007 "get_zone_info": false, 00:13:34.007 "zone_management": false, 00:13:34.007 "zone_append": false, 00:13:34.007 "compare": false, 00:13:34.007 "compare_and_write": false, 00:13:34.007 "abort": true, 00:13:34.007 "seek_hole": false, 00:13:34.007 "seek_data": false, 00:13:34.007 "copy": true, 00:13:34.007 "nvme_iov_md": false 00:13:34.007 }, 00:13:34.007 "memory_domains": [ 00:13:34.007 { 00:13:34.007 "dma_device_id": "system", 00:13:34.007 "dma_device_type": 1 00:13:34.007 }, 00:13:34.007 { 00:13:34.007 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:34.007 "dma_device_type": 2 00:13:34.007 } 00:13:34.007 ], 00:13:34.007 "driver_specific": {} 00:13:34.007 } 00:13:34.007 ] 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:34.007 "name": "Existed_Raid", 00:13:34.007 "uuid": "d77007c3-63fc-4fde-a2f8-d50add814a62", 00:13:34.007 "strip_size_kb": 64, 00:13:34.007 "state": "online", 00:13:34.007 "raid_level": "raid5f", 00:13:34.007 "superblock": false, 00:13:34.007 "num_base_bdevs": 3, 00:13:34.007 "num_base_bdevs_discovered": 3, 00:13:34.007 "num_base_bdevs_operational": 3, 00:13:34.007 "base_bdevs_list": [ 00:13:34.007 { 00:13:34.007 "name": "BaseBdev1", 00:13:34.007 "uuid": "fdf93e0d-63c3-4037-a351-e9185cc8eb77", 00:13:34.007 "is_configured": true, 00:13:34.007 "data_offset": 0, 00:13:34.007 "data_size": 65536 00:13:34.007 }, 00:13:34.007 { 00:13:34.007 "name": "BaseBdev2", 00:13:34.007 "uuid": "49fee062-8804-44ee-864b-ccf62b791e0a", 00:13:34.007 "is_configured": true, 00:13:34.007 "data_offset": 0, 00:13:34.007 "data_size": 65536 00:13:34.007 }, 00:13:34.007 { 00:13:34.007 "name": "BaseBdev3", 00:13:34.007 "uuid": "4815ca72-eec5-419e-9508-41cc45f5e7dc", 00:13:34.007 "is_configured": true, 00:13:34.007 "data_offset": 0, 00:13:34.007 "data_size": 65536 00:13:34.007 } 00:13:34.007 ] 00:13:34.007 }' 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:34.007 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:34.576 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:13:34.576 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:13:34.576 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:34.576 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:34.576 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:13:34.576 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:34.576 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:34.576 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:13:34.576 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.576 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:34.576 [2024-11-27 15:18:02.490372] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:34.576 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.576 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:34.576 "name": "Existed_Raid", 00:13:34.576 "aliases": [ 00:13:34.576 "d77007c3-63fc-4fde-a2f8-d50add814a62" 00:13:34.576 ], 00:13:34.576 "product_name": "Raid Volume", 00:13:34.576 "block_size": 512, 00:13:34.576 "num_blocks": 131072, 00:13:34.576 "uuid": "d77007c3-63fc-4fde-a2f8-d50add814a62", 00:13:34.576 "assigned_rate_limits": { 00:13:34.576 "rw_ios_per_sec": 0, 00:13:34.576 "rw_mbytes_per_sec": 0, 00:13:34.576 "r_mbytes_per_sec": 0, 00:13:34.576 "w_mbytes_per_sec": 0 00:13:34.576 }, 00:13:34.576 "claimed": false, 00:13:34.576 "zoned": false, 00:13:34.576 "supported_io_types": { 00:13:34.576 "read": true, 00:13:34.576 "write": true, 00:13:34.576 "unmap": false, 00:13:34.576 "flush": false, 00:13:34.576 "reset": true, 00:13:34.576 "nvme_admin": false, 00:13:34.576 "nvme_io": false, 00:13:34.576 "nvme_io_md": false, 00:13:34.576 "write_zeroes": true, 00:13:34.576 "zcopy": false, 00:13:34.576 "get_zone_info": false, 00:13:34.576 "zone_management": false, 00:13:34.576 "zone_append": false, 00:13:34.576 "compare": false, 00:13:34.576 "compare_and_write": false, 00:13:34.576 "abort": false, 00:13:34.576 "seek_hole": false, 00:13:34.576 "seek_data": false, 00:13:34.576 "copy": false, 00:13:34.576 "nvme_iov_md": false 00:13:34.576 }, 00:13:34.576 "driver_specific": { 00:13:34.576 "raid": { 00:13:34.576 "uuid": "d77007c3-63fc-4fde-a2f8-d50add814a62", 00:13:34.576 "strip_size_kb": 64, 00:13:34.576 "state": "online", 00:13:34.576 "raid_level": "raid5f", 00:13:34.576 "superblock": false, 00:13:34.576 "num_base_bdevs": 3, 00:13:34.576 "num_base_bdevs_discovered": 3, 00:13:34.576 "num_base_bdevs_operational": 3, 00:13:34.576 "base_bdevs_list": [ 00:13:34.576 { 00:13:34.576 "name": "BaseBdev1", 00:13:34.576 "uuid": "fdf93e0d-63c3-4037-a351-e9185cc8eb77", 00:13:34.577 "is_configured": true, 00:13:34.577 "data_offset": 0, 00:13:34.577 "data_size": 65536 00:13:34.577 }, 00:13:34.577 { 00:13:34.577 "name": "BaseBdev2", 00:13:34.577 "uuid": "49fee062-8804-44ee-864b-ccf62b791e0a", 00:13:34.577 "is_configured": true, 00:13:34.577 "data_offset": 0, 00:13:34.577 "data_size": 65536 00:13:34.577 }, 00:13:34.577 { 00:13:34.577 "name": "BaseBdev3", 00:13:34.577 "uuid": "4815ca72-eec5-419e-9508-41cc45f5e7dc", 00:13:34.577 "is_configured": true, 00:13:34.577 "data_offset": 0, 00:13:34.577 "data_size": 65536 00:13:34.577 } 00:13:34.577 ] 00:13:34.577 } 00:13:34.577 } 00:13:34.577 }' 00:13:34.577 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:34.577 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:13:34.577 BaseBdev2 00:13:34.577 BaseBdev3' 00:13:34.577 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:34.577 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:34.577 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:34.577 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:13:34.577 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.577 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:34.577 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:34.577 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.577 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:34.577 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:34.577 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:34.577 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:13:34.577 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.577 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:34.577 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:34.836 [2024-11-27 15:18:02.769709] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 2 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:34.836 "name": "Existed_Raid", 00:13:34.836 "uuid": "d77007c3-63fc-4fde-a2f8-d50add814a62", 00:13:34.836 "strip_size_kb": 64, 00:13:34.836 "state": "online", 00:13:34.836 "raid_level": "raid5f", 00:13:34.836 "superblock": false, 00:13:34.836 "num_base_bdevs": 3, 00:13:34.836 "num_base_bdevs_discovered": 2, 00:13:34.836 "num_base_bdevs_operational": 2, 00:13:34.836 "base_bdevs_list": [ 00:13:34.836 { 00:13:34.836 "name": null, 00:13:34.836 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:34.836 "is_configured": false, 00:13:34.836 "data_offset": 0, 00:13:34.836 "data_size": 65536 00:13:34.836 }, 00:13:34.836 { 00:13:34.836 "name": "BaseBdev2", 00:13:34.836 "uuid": "49fee062-8804-44ee-864b-ccf62b791e0a", 00:13:34.836 "is_configured": true, 00:13:34.836 "data_offset": 0, 00:13:34.836 "data_size": 65536 00:13:34.836 }, 00:13:34.836 { 00:13:34.836 "name": "BaseBdev3", 00:13:34.836 "uuid": "4815ca72-eec5-419e-9508-41cc45f5e7dc", 00:13:34.836 "is_configured": true, 00:13:34.836 "data_offset": 0, 00:13:34.836 "data_size": 65536 00:13:34.836 } 00:13:34.836 ] 00:13:34.836 }' 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:34.836 15:18:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.406 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:13:35.406 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.407 [2024-11-27 15:18:03.288513] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:35.407 [2024-11-27 15:18:03.288676] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:35.407 [2024-11-27 15:18:03.300014] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.407 [2024-11-27 15:18:03.355997] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:13:35.407 [2024-11-27 15:18:03.356095] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.407 BaseBdev2 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.407 [ 00:13:35.407 { 00:13:35.407 "name": "BaseBdev2", 00:13:35.407 "aliases": [ 00:13:35.407 "9b3daaf4-0226-4834-b8f0-d531e22ec559" 00:13:35.407 ], 00:13:35.407 "product_name": "Malloc disk", 00:13:35.407 "block_size": 512, 00:13:35.407 "num_blocks": 65536, 00:13:35.407 "uuid": "9b3daaf4-0226-4834-b8f0-d531e22ec559", 00:13:35.407 "assigned_rate_limits": { 00:13:35.407 "rw_ios_per_sec": 0, 00:13:35.407 "rw_mbytes_per_sec": 0, 00:13:35.407 "r_mbytes_per_sec": 0, 00:13:35.407 "w_mbytes_per_sec": 0 00:13:35.407 }, 00:13:35.407 "claimed": false, 00:13:35.407 "zoned": false, 00:13:35.407 "supported_io_types": { 00:13:35.407 "read": true, 00:13:35.407 "write": true, 00:13:35.407 "unmap": true, 00:13:35.407 "flush": true, 00:13:35.407 "reset": true, 00:13:35.407 "nvme_admin": false, 00:13:35.407 "nvme_io": false, 00:13:35.407 "nvme_io_md": false, 00:13:35.407 "write_zeroes": true, 00:13:35.407 "zcopy": true, 00:13:35.407 "get_zone_info": false, 00:13:35.407 "zone_management": false, 00:13:35.407 "zone_append": false, 00:13:35.407 "compare": false, 00:13:35.407 "compare_and_write": false, 00:13:35.407 "abort": true, 00:13:35.407 "seek_hole": false, 00:13:35.407 "seek_data": false, 00:13:35.407 "copy": true, 00:13:35.407 "nvme_iov_md": false 00:13:35.407 }, 00:13:35.407 "memory_domains": [ 00:13:35.407 { 00:13:35.407 "dma_device_id": "system", 00:13:35.407 "dma_device_type": 1 00:13:35.407 }, 00:13:35.407 { 00:13:35.407 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:35.407 "dma_device_type": 2 00:13:35.407 } 00:13:35.407 ], 00:13:35.407 "driver_specific": {} 00:13:35.407 } 00:13:35.407 ] 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.407 BaseBdev3 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.407 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.667 [ 00:13:35.667 { 00:13:35.667 "name": "BaseBdev3", 00:13:35.667 "aliases": [ 00:13:35.667 "6c390a5c-dfe7-4024-b5bd-fde7a4b87cc4" 00:13:35.667 ], 00:13:35.667 "product_name": "Malloc disk", 00:13:35.667 "block_size": 512, 00:13:35.667 "num_blocks": 65536, 00:13:35.667 "uuid": "6c390a5c-dfe7-4024-b5bd-fde7a4b87cc4", 00:13:35.667 "assigned_rate_limits": { 00:13:35.667 "rw_ios_per_sec": 0, 00:13:35.667 "rw_mbytes_per_sec": 0, 00:13:35.667 "r_mbytes_per_sec": 0, 00:13:35.667 "w_mbytes_per_sec": 0 00:13:35.667 }, 00:13:35.667 "claimed": false, 00:13:35.667 "zoned": false, 00:13:35.667 "supported_io_types": { 00:13:35.667 "read": true, 00:13:35.667 "write": true, 00:13:35.667 "unmap": true, 00:13:35.667 "flush": true, 00:13:35.667 "reset": true, 00:13:35.667 "nvme_admin": false, 00:13:35.667 "nvme_io": false, 00:13:35.667 "nvme_io_md": false, 00:13:35.667 "write_zeroes": true, 00:13:35.667 "zcopy": true, 00:13:35.667 "get_zone_info": false, 00:13:35.667 "zone_management": false, 00:13:35.667 "zone_append": false, 00:13:35.667 "compare": false, 00:13:35.667 "compare_and_write": false, 00:13:35.667 "abort": true, 00:13:35.667 "seek_hole": false, 00:13:35.667 "seek_data": false, 00:13:35.667 "copy": true, 00:13:35.667 "nvme_iov_md": false 00:13:35.667 }, 00:13:35.667 "memory_domains": [ 00:13:35.667 { 00:13:35.667 "dma_device_id": "system", 00:13:35.667 "dma_device_type": 1 00:13:35.667 }, 00:13:35.667 { 00:13:35.667 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:35.667 "dma_device_type": 2 00:13:35.667 } 00:13:35.667 ], 00:13:35.667 "driver_specific": {} 00:13:35.667 } 00:13:35.667 ] 00:13:35.667 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.667 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:13:35.667 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:13:35.667 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:35.667 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:35.667 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.667 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.667 [2024-11-27 15:18:03.533181] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:35.667 [2024-11-27 15:18:03.533280] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:35.667 [2024-11-27 15:18:03.533320] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:35.667 [2024-11-27 15:18:03.535105] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:35.667 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.667 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:35.667 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:35.668 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:35.668 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:35.668 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:35.668 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:35.668 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:35.668 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:35.668 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:35.668 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:35.668 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:35.668 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.668 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:35.668 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.668 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.668 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:35.668 "name": "Existed_Raid", 00:13:35.668 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:35.668 "strip_size_kb": 64, 00:13:35.668 "state": "configuring", 00:13:35.668 "raid_level": "raid5f", 00:13:35.668 "superblock": false, 00:13:35.668 "num_base_bdevs": 3, 00:13:35.668 "num_base_bdevs_discovered": 2, 00:13:35.668 "num_base_bdevs_operational": 3, 00:13:35.668 "base_bdevs_list": [ 00:13:35.668 { 00:13:35.668 "name": "BaseBdev1", 00:13:35.668 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:35.668 "is_configured": false, 00:13:35.668 "data_offset": 0, 00:13:35.668 "data_size": 0 00:13:35.668 }, 00:13:35.668 { 00:13:35.668 "name": "BaseBdev2", 00:13:35.668 "uuid": "9b3daaf4-0226-4834-b8f0-d531e22ec559", 00:13:35.668 "is_configured": true, 00:13:35.668 "data_offset": 0, 00:13:35.668 "data_size": 65536 00:13:35.668 }, 00:13:35.668 { 00:13:35.668 "name": "BaseBdev3", 00:13:35.668 "uuid": "6c390a5c-dfe7-4024-b5bd-fde7a4b87cc4", 00:13:35.668 "is_configured": true, 00:13:35.668 "data_offset": 0, 00:13:35.668 "data_size": 65536 00:13:35.668 } 00:13:35.668 ] 00:13:35.668 }' 00:13:35.668 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:35.668 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.928 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:13:35.928 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.928 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.928 [2024-11-27 15:18:03.964448] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:35.928 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.928 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:35.928 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:35.928 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:35.928 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:35.928 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:35.928 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:35.928 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:35.928 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:35.928 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:35.928 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:35.928 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:35.928 15:18:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:35.928 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.928 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.928 15:18:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.928 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:35.928 "name": "Existed_Raid", 00:13:35.928 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:35.928 "strip_size_kb": 64, 00:13:35.928 "state": "configuring", 00:13:35.928 "raid_level": "raid5f", 00:13:35.928 "superblock": false, 00:13:35.928 "num_base_bdevs": 3, 00:13:35.928 "num_base_bdevs_discovered": 1, 00:13:35.928 "num_base_bdevs_operational": 3, 00:13:35.928 "base_bdevs_list": [ 00:13:35.928 { 00:13:35.928 "name": "BaseBdev1", 00:13:35.928 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:35.928 "is_configured": false, 00:13:35.928 "data_offset": 0, 00:13:35.928 "data_size": 0 00:13:35.928 }, 00:13:35.928 { 00:13:35.928 "name": null, 00:13:35.928 "uuid": "9b3daaf4-0226-4834-b8f0-d531e22ec559", 00:13:35.928 "is_configured": false, 00:13:35.928 "data_offset": 0, 00:13:35.928 "data_size": 65536 00:13:35.928 }, 00:13:35.928 { 00:13:35.928 "name": "BaseBdev3", 00:13:35.928 "uuid": "6c390a5c-dfe7-4024-b5bd-fde7a4b87cc4", 00:13:35.928 "is_configured": true, 00:13:35.928 "data_offset": 0, 00:13:35.928 "data_size": 65536 00:13:35.928 } 00:13:35.928 ] 00:13:35.928 }' 00:13:35.928 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:35.928 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.497 [2024-11-27 15:18:04.422853] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:36.497 BaseBdev1 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.497 [ 00:13:36.497 { 00:13:36.497 "name": "BaseBdev1", 00:13:36.497 "aliases": [ 00:13:36.497 "e593647f-eab4-48dc-8248-c4494492dfcd" 00:13:36.497 ], 00:13:36.497 "product_name": "Malloc disk", 00:13:36.497 "block_size": 512, 00:13:36.497 "num_blocks": 65536, 00:13:36.497 "uuid": "e593647f-eab4-48dc-8248-c4494492dfcd", 00:13:36.497 "assigned_rate_limits": { 00:13:36.497 "rw_ios_per_sec": 0, 00:13:36.497 "rw_mbytes_per_sec": 0, 00:13:36.497 "r_mbytes_per_sec": 0, 00:13:36.497 "w_mbytes_per_sec": 0 00:13:36.497 }, 00:13:36.497 "claimed": true, 00:13:36.497 "claim_type": "exclusive_write", 00:13:36.497 "zoned": false, 00:13:36.497 "supported_io_types": { 00:13:36.497 "read": true, 00:13:36.497 "write": true, 00:13:36.497 "unmap": true, 00:13:36.497 "flush": true, 00:13:36.497 "reset": true, 00:13:36.497 "nvme_admin": false, 00:13:36.497 "nvme_io": false, 00:13:36.497 "nvme_io_md": false, 00:13:36.497 "write_zeroes": true, 00:13:36.497 "zcopy": true, 00:13:36.497 "get_zone_info": false, 00:13:36.497 "zone_management": false, 00:13:36.497 "zone_append": false, 00:13:36.497 "compare": false, 00:13:36.497 "compare_and_write": false, 00:13:36.497 "abort": true, 00:13:36.497 "seek_hole": false, 00:13:36.497 "seek_data": false, 00:13:36.497 "copy": true, 00:13:36.497 "nvme_iov_md": false 00:13:36.497 }, 00:13:36.497 "memory_domains": [ 00:13:36.497 { 00:13:36.497 "dma_device_id": "system", 00:13:36.497 "dma_device_type": 1 00:13:36.497 }, 00:13:36.497 { 00:13:36.497 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:36.497 "dma_device_type": 2 00:13:36.497 } 00:13:36.497 ], 00:13:36.497 "driver_specific": {} 00:13:36.497 } 00:13:36.497 ] 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:36.497 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:36.497 "name": "Existed_Raid", 00:13:36.497 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:36.497 "strip_size_kb": 64, 00:13:36.497 "state": "configuring", 00:13:36.497 "raid_level": "raid5f", 00:13:36.497 "superblock": false, 00:13:36.497 "num_base_bdevs": 3, 00:13:36.497 "num_base_bdevs_discovered": 2, 00:13:36.497 "num_base_bdevs_operational": 3, 00:13:36.497 "base_bdevs_list": [ 00:13:36.497 { 00:13:36.497 "name": "BaseBdev1", 00:13:36.497 "uuid": "e593647f-eab4-48dc-8248-c4494492dfcd", 00:13:36.497 "is_configured": true, 00:13:36.497 "data_offset": 0, 00:13:36.497 "data_size": 65536 00:13:36.497 }, 00:13:36.497 { 00:13:36.497 "name": null, 00:13:36.497 "uuid": "9b3daaf4-0226-4834-b8f0-d531e22ec559", 00:13:36.497 "is_configured": false, 00:13:36.497 "data_offset": 0, 00:13:36.497 "data_size": 65536 00:13:36.497 }, 00:13:36.497 { 00:13:36.497 "name": "BaseBdev3", 00:13:36.497 "uuid": "6c390a5c-dfe7-4024-b5bd-fde7a4b87cc4", 00:13:36.497 "is_configured": true, 00:13:36.497 "data_offset": 0, 00:13:36.497 "data_size": 65536 00:13:36.497 } 00:13:36.498 ] 00:13:36.498 }' 00:13:36.498 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:36.498 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.067 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:37.067 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.067 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.067 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:13:37.067 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.067 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:13:37.067 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:13:37.067 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.067 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.067 [2024-11-27 15:18:04.962004] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:13:37.067 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.067 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:37.067 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:37.067 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:37.067 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:37.067 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:37.067 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:37.067 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:37.067 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:37.067 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:37.067 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:37.067 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:37.067 15:18:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:37.067 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.067 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.067 15:18:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.067 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:37.067 "name": "Existed_Raid", 00:13:37.067 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:37.067 "strip_size_kb": 64, 00:13:37.067 "state": "configuring", 00:13:37.067 "raid_level": "raid5f", 00:13:37.067 "superblock": false, 00:13:37.067 "num_base_bdevs": 3, 00:13:37.067 "num_base_bdevs_discovered": 1, 00:13:37.067 "num_base_bdevs_operational": 3, 00:13:37.067 "base_bdevs_list": [ 00:13:37.067 { 00:13:37.067 "name": "BaseBdev1", 00:13:37.067 "uuid": "e593647f-eab4-48dc-8248-c4494492dfcd", 00:13:37.067 "is_configured": true, 00:13:37.067 "data_offset": 0, 00:13:37.067 "data_size": 65536 00:13:37.067 }, 00:13:37.067 { 00:13:37.067 "name": null, 00:13:37.067 "uuid": "9b3daaf4-0226-4834-b8f0-d531e22ec559", 00:13:37.068 "is_configured": false, 00:13:37.068 "data_offset": 0, 00:13:37.068 "data_size": 65536 00:13:37.068 }, 00:13:37.068 { 00:13:37.068 "name": null, 00:13:37.068 "uuid": "6c390a5c-dfe7-4024-b5bd-fde7a4b87cc4", 00:13:37.068 "is_configured": false, 00:13:37.068 "data_offset": 0, 00:13:37.068 "data_size": 65536 00:13:37.068 } 00:13:37.068 ] 00:13:37.068 }' 00:13:37.068 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:37.068 15:18:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.327 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:37.327 15:18:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.327 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:13:37.327 15:18:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.327 15:18:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.586 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:13:37.586 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:13:37.586 15:18:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.586 15:18:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.587 [2024-11-27 15:18:05.441161] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:37.587 15:18:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.587 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:37.587 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:37.587 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:37.587 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:37.587 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:37.587 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:37.587 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:37.587 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:37.587 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:37.587 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:37.587 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:37.587 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:37.587 15:18:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.587 15:18:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.587 15:18:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.587 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:37.587 "name": "Existed_Raid", 00:13:37.587 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:37.587 "strip_size_kb": 64, 00:13:37.587 "state": "configuring", 00:13:37.587 "raid_level": "raid5f", 00:13:37.587 "superblock": false, 00:13:37.587 "num_base_bdevs": 3, 00:13:37.587 "num_base_bdevs_discovered": 2, 00:13:37.587 "num_base_bdevs_operational": 3, 00:13:37.587 "base_bdevs_list": [ 00:13:37.587 { 00:13:37.587 "name": "BaseBdev1", 00:13:37.587 "uuid": "e593647f-eab4-48dc-8248-c4494492dfcd", 00:13:37.587 "is_configured": true, 00:13:37.587 "data_offset": 0, 00:13:37.587 "data_size": 65536 00:13:37.587 }, 00:13:37.587 { 00:13:37.587 "name": null, 00:13:37.587 "uuid": "9b3daaf4-0226-4834-b8f0-d531e22ec559", 00:13:37.587 "is_configured": false, 00:13:37.587 "data_offset": 0, 00:13:37.587 "data_size": 65536 00:13:37.587 }, 00:13:37.587 { 00:13:37.587 "name": "BaseBdev3", 00:13:37.587 "uuid": "6c390a5c-dfe7-4024-b5bd-fde7a4b87cc4", 00:13:37.587 "is_configured": true, 00:13:37.587 "data_offset": 0, 00:13:37.587 "data_size": 65536 00:13:37.587 } 00:13:37.587 ] 00:13:37.587 }' 00:13:37.587 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:37.587 15:18:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.846 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:37.846 15:18:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.846 15:18:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.846 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:13:37.846 15:18:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.846 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:13:37.846 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:13:37.846 15:18:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.846 15:18:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.846 [2024-11-27 15:18:05.928392] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:37.846 15:18:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.846 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:37.846 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:37.846 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:37.846 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:37.846 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:37.846 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:37.846 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:37.846 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:37.846 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:37.846 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:37.846 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:37.846 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:37.846 15:18:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.846 15:18:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:38.105 15:18:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:38.105 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:38.105 "name": "Existed_Raid", 00:13:38.105 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:38.105 "strip_size_kb": 64, 00:13:38.105 "state": "configuring", 00:13:38.105 "raid_level": "raid5f", 00:13:38.105 "superblock": false, 00:13:38.105 "num_base_bdevs": 3, 00:13:38.105 "num_base_bdevs_discovered": 1, 00:13:38.105 "num_base_bdevs_operational": 3, 00:13:38.105 "base_bdevs_list": [ 00:13:38.105 { 00:13:38.105 "name": null, 00:13:38.105 "uuid": "e593647f-eab4-48dc-8248-c4494492dfcd", 00:13:38.105 "is_configured": false, 00:13:38.105 "data_offset": 0, 00:13:38.105 "data_size": 65536 00:13:38.105 }, 00:13:38.105 { 00:13:38.105 "name": null, 00:13:38.105 "uuid": "9b3daaf4-0226-4834-b8f0-d531e22ec559", 00:13:38.105 "is_configured": false, 00:13:38.105 "data_offset": 0, 00:13:38.105 "data_size": 65536 00:13:38.105 }, 00:13:38.105 { 00:13:38.105 "name": "BaseBdev3", 00:13:38.105 "uuid": "6c390a5c-dfe7-4024-b5bd-fde7a4b87cc4", 00:13:38.105 "is_configured": true, 00:13:38.105 "data_offset": 0, 00:13:38.105 "data_size": 65536 00:13:38.105 } 00:13:38.105 ] 00:13:38.105 }' 00:13:38.105 15:18:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:38.105 15:18:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:38.364 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:13:38.364 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:38.364 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:38.364 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:38.364 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:38.364 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:13:38.364 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:13:38.364 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:38.364 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:38.364 [2024-11-27 15:18:06.418259] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:38.364 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:38.364 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:38.364 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:38.364 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:38.364 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:38.364 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:38.364 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:38.364 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:38.364 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:38.364 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:38.364 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:38.364 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:38.364 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:38.364 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:38.364 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:38.364 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:38.623 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:38.623 "name": "Existed_Raid", 00:13:38.623 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:38.623 "strip_size_kb": 64, 00:13:38.623 "state": "configuring", 00:13:38.623 "raid_level": "raid5f", 00:13:38.624 "superblock": false, 00:13:38.624 "num_base_bdevs": 3, 00:13:38.624 "num_base_bdevs_discovered": 2, 00:13:38.624 "num_base_bdevs_operational": 3, 00:13:38.624 "base_bdevs_list": [ 00:13:38.624 { 00:13:38.624 "name": null, 00:13:38.624 "uuid": "e593647f-eab4-48dc-8248-c4494492dfcd", 00:13:38.624 "is_configured": false, 00:13:38.624 "data_offset": 0, 00:13:38.624 "data_size": 65536 00:13:38.624 }, 00:13:38.624 { 00:13:38.624 "name": "BaseBdev2", 00:13:38.624 "uuid": "9b3daaf4-0226-4834-b8f0-d531e22ec559", 00:13:38.624 "is_configured": true, 00:13:38.624 "data_offset": 0, 00:13:38.624 "data_size": 65536 00:13:38.624 }, 00:13:38.624 { 00:13:38.624 "name": "BaseBdev3", 00:13:38.624 "uuid": "6c390a5c-dfe7-4024-b5bd-fde7a4b87cc4", 00:13:38.624 "is_configured": true, 00:13:38.624 "data_offset": 0, 00:13:38.624 "data_size": 65536 00:13:38.624 } 00:13:38.624 ] 00:13:38.624 }' 00:13:38.624 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:38.624 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:38.883 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:13:38.883 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:38.883 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:38.883 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:38.883 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:38.883 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:13:38.883 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:38.883 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:13:38.883 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:38.883 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:38.883 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:38.883 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u e593647f-eab4-48dc-8248-c4494492dfcd 00:13:38.883 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:38.883 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:38.883 [2024-11-27 15:18:06.952760] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:13:38.883 [2024-11-27 15:18:06.952870] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:13:38.883 [2024-11-27 15:18:06.952910] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:13:38.883 [2024-11-27 15:18:06.953219] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:13:38.883 [2024-11-27 15:18:06.953684] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:13:38.883 NewBaseBdev 00:13:38.883 [2024-11-27 15:18:06.953733] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006d00 00:13:38.883 [2024-11-27 15:18:06.953942] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:38.883 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:38.883 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:13:38.883 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:13:38.883 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:38.883 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:13:38.883 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:38.883 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:38.883 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:38.883 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:38.883 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:38.883 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:38.883 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:13:38.883 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:38.883 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:38.883 [ 00:13:38.883 { 00:13:38.883 "name": "NewBaseBdev", 00:13:38.883 "aliases": [ 00:13:38.883 "e593647f-eab4-48dc-8248-c4494492dfcd" 00:13:38.883 ], 00:13:38.883 "product_name": "Malloc disk", 00:13:38.883 "block_size": 512, 00:13:38.883 "num_blocks": 65536, 00:13:38.883 "uuid": "e593647f-eab4-48dc-8248-c4494492dfcd", 00:13:38.883 "assigned_rate_limits": { 00:13:38.883 "rw_ios_per_sec": 0, 00:13:38.883 "rw_mbytes_per_sec": 0, 00:13:38.883 "r_mbytes_per_sec": 0, 00:13:38.883 "w_mbytes_per_sec": 0 00:13:38.883 }, 00:13:38.883 "claimed": true, 00:13:38.883 "claim_type": "exclusive_write", 00:13:38.883 "zoned": false, 00:13:39.143 "supported_io_types": { 00:13:39.143 "read": true, 00:13:39.143 "write": true, 00:13:39.143 "unmap": true, 00:13:39.143 "flush": true, 00:13:39.143 "reset": true, 00:13:39.143 "nvme_admin": false, 00:13:39.143 "nvme_io": false, 00:13:39.143 "nvme_io_md": false, 00:13:39.143 "write_zeroes": true, 00:13:39.143 "zcopy": true, 00:13:39.143 "get_zone_info": false, 00:13:39.143 "zone_management": false, 00:13:39.143 "zone_append": false, 00:13:39.143 "compare": false, 00:13:39.143 "compare_and_write": false, 00:13:39.143 "abort": true, 00:13:39.143 "seek_hole": false, 00:13:39.143 "seek_data": false, 00:13:39.143 "copy": true, 00:13:39.143 "nvme_iov_md": false 00:13:39.143 }, 00:13:39.143 "memory_domains": [ 00:13:39.143 { 00:13:39.143 "dma_device_id": "system", 00:13:39.143 "dma_device_type": 1 00:13:39.143 }, 00:13:39.143 { 00:13:39.143 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:39.143 "dma_device_type": 2 00:13:39.143 } 00:13:39.143 ], 00:13:39.143 "driver_specific": {} 00:13:39.143 } 00:13:39.143 ] 00:13:39.143 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:39.143 15:18:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:13:39.143 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:13:39.143 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:39.143 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:39.143 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:39.143 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:39.143 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:39.143 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:39.143 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:39.143 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:39.143 15:18:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:39.143 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:39.143 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:39.143 15:18:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:39.143 15:18:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:39.143 15:18:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:39.143 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:39.143 "name": "Existed_Raid", 00:13:39.143 "uuid": "7655a41b-0ef2-4af2-bb97-b8d093f3d968", 00:13:39.143 "strip_size_kb": 64, 00:13:39.143 "state": "online", 00:13:39.143 "raid_level": "raid5f", 00:13:39.143 "superblock": false, 00:13:39.143 "num_base_bdevs": 3, 00:13:39.143 "num_base_bdevs_discovered": 3, 00:13:39.143 "num_base_bdevs_operational": 3, 00:13:39.143 "base_bdevs_list": [ 00:13:39.143 { 00:13:39.143 "name": "NewBaseBdev", 00:13:39.143 "uuid": "e593647f-eab4-48dc-8248-c4494492dfcd", 00:13:39.143 "is_configured": true, 00:13:39.143 "data_offset": 0, 00:13:39.143 "data_size": 65536 00:13:39.143 }, 00:13:39.143 { 00:13:39.143 "name": "BaseBdev2", 00:13:39.143 "uuid": "9b3daaf4-0226-4834-b8f0-d531e22ec559", 00:13:39.143 "is_configured": true, 00:13:39.143 "data_offset": 0, 00:13:39.143 "data_size": 65536 00:13:39.143 }, 00:13:39.143 { 00:13:39.143 "name": "BaseBdev3", 00:13:39.143 "uuid": "6c390a5c-dfe7-4024-b5bd-fde7a4b87cc4", 00:13:39.143 "is_configured": true, 00:13:39.143 "data_offset": 0, 00:13:39.143 "data_size": 65536 00:13:39.143 } 00:13:39.143 ] 00:13:39.143 }' 00:13:39.143 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:39.143 15:18:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:39.402 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:13:39.402 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:13:39.402 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:39.402 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:39.402 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:13:39.402 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:39.402 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:39.402 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:13:39.402 15:18:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:39.402 15:18:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:39.402 [2024-11-27 15:18:07.412236] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:39.402 15:18:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:39.402 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:39.402 "name": "Existed_Raid", 00:13:39.402 "aliases": [ 00:13:39.402 "7655a41b-0ef2-4af2-bb97-b8d093f3d968" 00:13:39.402 ], 00:13:39.402 "product_name": "Raid Volume", 00:13:39.402 "block_size": 512, 00:13:39.402 "num_blocks": 131072, 00:13:39.402 "uuid": "7655a41b-0ef2-4af2-bb97-b8d093f3d968", 00:13:39.402 "assigned_rate_limits": { 00:13:39.402 "rw_ios_per_sec": 0, 00:13:39.402 "rw_mbytes_per_sec": 0, 00:13:39.402 "r_mbytes_per_sec": 0, 00:13:39.402 "w_mbytes_per_sec": 0 00:13:39.402 }, 00:13:39.402 "claimed": false, 00:13:39.402 "zoned": false, 00:13:39.402 "supported_io_types": { 00:13:39.402 "read": true, 00:13:39.402 "write": true, 00:13:39.402 "unmap": false, 00:13:39.402 "flush": false, 00:13:39.402 "reset": true, 00:13:39.402 "nvme_admin": false, 00:13:39.402 "nvme_io": false, 00:13:39.402 "nvme_io_md": false, 00:13:39.402 "write_zeroes": true, 00:13:39.402 "zcopy": false, 00:13:39.402 "get_zone_info": false, 00:13:39.402 "zone_management": false, 00:13:39.402 "zone_append": false, 00:13:39.402 "compare": false, 00:13:39.402 "compare_and_write": false, 00:13:39.402 "abort": false, 00:13:39.402 "seek_hole": false, 00:13:39.402 "seek_data": false, 00:13:39.402 "copy": false, 00:13:39.402 "nvme_iov_md": false 00:13:39.402 }, 00:13:39.402 "driver_specific": { 00:13:39.402 "raid": { 00:13:39.402 "uuid": "7655a41b-0ef2-4af2-bb97-b8d093f3d968", 00:13:39.402 "strip_size_kb": 64, 00:13:39.402 "state": "online", 00:13:39.402 "raid_level": "raid5f", 00:13:39.402 "superblock": false, 00:13:39.402 "num_base_bdevs": 3, 00:13:39.402 "num_base_bdevs_discovered": 3, 00:13:39.402 "num_base_bdevs_operational": 3, 00:13:39.402 "base_bdevs_list": [ 00:13:39.402 { 00:13:39.402 "name": "NewBaseBdev", 00:13:39.402 "uuid": "e593647f-eab4-48dc-8248-c4494492dfcd", 00:13:39.402 "is_configured": true, 00:13:39.402 "data_offset": 0, 00:13:39.402 "data_size": 65536 00:13:39.402 }, 00:13:39.402 { 00:13:39.402 "name": "BaseBdev2", 00:13:39.402 "uuid": "9b3daaf4-0226-4834-b8f0-d531e22ec559", 00:13:39.402 "is_configured": true, 00:13:39.402 "data_offset": 0, 00:13:39.402 "data_size": 65536 00:13:39.402 }, 00:13:39.402 { 00:13:39.402 "name": "BaseBdev3", 00:13:39.402 "uuid": "6c390a5c-dfe7-4024-b5bd-fde7a4b87cc4", 00:13:39.402 "is_configured": true, 00:13:39.402 "data_offset": 0, 00:13:39.402 "data_size": 65536 00:13:39.402 } 00:13:39.402 ] 00:13:39.402 } 00:13:39.402 } 00:13:39.402 }' 00:13:39.402 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:39.402 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:13:39.402 BaseBdev2 00:13:39.402 BaseBdev3' 00:13:39.402 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:39.662 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:39.662 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:39.662 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:13:39.662 15:18:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:39.662 15:18:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:39.662 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:39.662 15:18:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:39.662 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:39.662 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:39.662 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:39.662 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:39.662 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:13:39.662 15:18:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:39.662 15:18:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:39.662 15:18:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:39.662 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:39.662 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:39.662 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:39.662 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:13:39.662 15:18:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:39.662 15:18:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:39.662 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:39.662 15:18:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:39.662 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:39.662 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:39.663 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:39.663 15:18:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:39.663 15:18:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:39.663 [2024-11-27 15:18:07.683584] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:39.663 [2024-11-27 15:18:07.683660] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:39.663 [2024-11-27 15:18:07.683773] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:39.663 [2024-11-27 15:18:07.684067] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:39.663 [2024-11-27 15:18:07.684130] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name Existed_Raid, state offline 00:13:39.663 15:18:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:39.663 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 90634 00:13:39.663 15:18:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 90634 ']' 00:13:39.663 15:18:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@958 -- # kill -0 90634 00:13:39.663 15:18:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@959 -- # uname 00:13:39.663 15:18:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:39.663 15:18:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 90634 00:13:39.663 killing process with pid 90634 00:13:39.663 15:18:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:39.663 15:18:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:39.663 15:18:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 90634' 00:13:39.663 15:18:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@973 -- # kill 90634 00:13:39.663 [2024-11-27 15:18:07.719153] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:39.663 15:18:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@978 -- # wait 90634 00:13:39.663 [2024-11-27 15:18:07.751032] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:39.923 15:18:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:13:39.923 ************************************ 00:13:39.923 END TEST raid5f_state_function_test 00:13:39.923 ************************************ 00:13:39.923 00:13:39.923 real 0m8.862s 00:13:39.923 user 0m15.135s 00:13:39.923 sys 0m1.859s 00:13:39.923 15:18:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:39.923 15:18:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:39.923 15:18:08 bdev_raid -- bdev/bdev_raid.sh@987 -- # run_test raid5f_state_function_test_sb raid_state_function_test raid5f 3 true 00:13:39.923 15:18:08 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:13:39.923 15:18:08 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:39.923 15:18:08 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:40.183 ************************************ 00:13:40.183 START TEST raid5f_state_function_test_sb 00:13:40.183 ************************************ 00:13:40.183 15:18:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid5f 3 true 00:13:40.183 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=91239 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 91239' 00:13:40.184 Process raid pid: 91239 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 91239 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 91239 ']' 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:40.184 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:13:40.184 15:18:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:40.184 [2024-11-27 15:18:08.125322] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:13:40.184 [2024-11-27 15:18:08.125456] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:40.443 [2024-11-27 15:18:08.298189] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:40.443 [2024-11-27 15:18:08.328687] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:40.443 [2024-11-27 15:18:08.372340] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:40.443 [2024-11-27 15:18:08.372461] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:41.012 15:18:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:13:41.012 15:18:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:13:41.012 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:41.012 15:18:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.012 15:18:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:41.012 [2024-11-27 15:18:08.961010] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:41.012 [2024-11-27 15:18:08.961145] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:41.012 [2024-11-27 15:18:08.961175] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:41.012 [2024-11-27 15:18:08.961199] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:41.012 [2024-11-27 15:18:08.961220] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:41.012 [2024-11-27 15:18:08.961246] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:41.012 15:18:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.012 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:41.012 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:41.012 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:41.012 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:41.012 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:41.012 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:41.012 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:41.012 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:41.012 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:41.012 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:41.012 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:41.012 15:18:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:41.012 15:18:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.012 15:18:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:41.012 15:18:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.012 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:41.012 "name": "Existed_Raid", 00:13:41.012 "uuid": "32ddab81-ec62-43e8-b843-07de94198f76", 00:13:41.012 "strip_size_kb": 64, 00:13:41.012 "state": "configuring", 00:13:41.012 "raid_level": "raid5f", 00:13:41.012 "superblock": true, 00:13:41.012 "num_base_bdevs": 3, 00:13:41.012 "num_base_bdevs_discovered": 0, 00:13:41.012 "num_base_bdevs_operational": 3, 00:13:41.012 "base_bdevs_list": [ 00:13:41.012 { 00:13:41.012 "name": "BaseBdev1", 00:13:41.012 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:41.012 "is_configured": false, 00:13:41.012 "data_offset": 0, 00:13:41.012 "data_size": 0 00:13:41.012 }, 00:13:41.012 { 00:13:41.012 "name": "BaseBdev2", 00:13:41.012 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:41.012 "is_configured": false, 00:13:41.012 "data_offset": 0, 00:13:41.012 "data_size": 0 00:13:41.012 }, 00:13:41.012 { 00:13:41.012 "name": "BaseBdev3", 00:13:41.012 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:41.012 "is_configured": false, 00:13:41.012 "data_offset": 0, 00:13:41.012 "data_size": 0 00:13:41.012 } 00:13:41.012 ] 00:13:41.012 }' 00:13:41.012 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:41.012 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:41.580 [2024-11-27 15:18:09.424078] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:41.580 [2024-11-27 15:18:09.424167] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:41.580 [2024-11-27 15:18:09.436073] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:41.580 [2024-11-27 15:18:09.436163] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:41.580 [2024-11-27 15:18:09.436175] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:41.580 [2024-11-27 15:18:09.436184] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:41.580 [2024-11-27 15:18:09.436190] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:41.580 [2024-11-27 15:18:09.436199] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:41.580 [2024-11-27 15:18:09.457178] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:41.580 BaseBdev1 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:41.580 [ 00:13:41.580 { 00:13:41.580 "name": "BaseBdev1", 00:13:41.580 "aliases": [ 00:13:41.580 "d97d609e-aef4-4a54-aaa2-2ff418116f8b" 00:13:41.580 ], 00:13:41.580 "product_name": "Malloc disk", 00:13:41.580 "block_size": 512, 00:13:41.580 "num_blocks": 65536, 00:13:41.580 "uuid": "d97d609e-aef4-4a54-aaa2-2ff418116f8b", 00:13:41.580 "assigned_rate_limits": { 00:13:41.580 "rw_ios_per_sec": 0, 00:13:41.580 "rw_mbytes_per_sec": 0, 00:13:41.580 "r_mbytes_per_sec": 0, 00:13:41.580 "w_mbytes_per_sec": 0 00:13:41.580 }, 00:13:41.580 "claimed": true, 00:13:41.580 "claim_type": "exclusive_write", 00:13:41.580 "zoned": false, 00:13:41.580 "supported_io_types": { 00:13:41.580 "read": true, 00:13:41.580 "write": true, 00:13:41.580 "unmap": true, 00:13:41.580 "flush": true, 00:13:41.580 "reset": true, 00:13:41.580 "nvme_admin": false, 00:13:41.580 "nvme_io": false, 00:13:41.580 "nvme_io_md": false, 00:13:41.580 "write_zeroes": true, 00:13:41.580 "zcopy": true, 00:13:41.580 "get_zone_info": false, 00:13:41.580 "zone_management": false, 00:13:41.580 "zone_append": false, 00:13:41.580 "compare": false, 00:13:41.580 "compare_and_write": false, 00:13:41.580 "abort": true, 00:13:41.580 "seek_hole": false, 00:13:41.580 "seek_data": false, 00:13:41.580 "copy": true, 00:13:41.580 "nvme_iov_md": false 00:13:41.580 }, 00:13:41.580 "memory_domains": [ 00:13:41.580 { 00:13:41.580 "dma_device_id": "system", 00:13:41.580 "dma_device_type": 1 00:13:41.580 }, 00:13:41.580 { 00:13:41.580 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:41.580 "dma_device_type": 2 00:13:41.580 } 00:13:41.580 ], 00:13:41.580 "driver_specific": {} 00:13:41.580 } 00:13:41.580 ] 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:41.580 "name": "Existed_Raid", 00:13:41.580 "uuid": "78585281-2518-41e0-9a7f-0da7b01213d3", 00:13:41.580 "strip_size_kb": 64, 00:13:41.580 "state": "configuring", 00:13:41.580 "raid_level": "raid5f", 00:13:41.580 "superblock": true, 00:13:41.580 "num_base_bdevs": 3, 00:13:41.580 "num_base_bdevs_discovered": 1, 00:13:41.580 "num_base_bdevs_operational": 3, 00:13:41.580 "base_bdevs_list": [ 00:13:41.580 { 00:13:41.580 "name": "BaseBdev1", 00:13:41.580 "uuid": "d97d609e-aef4-4a54-aaa2-2ff418116f8b", 00:13:41.580 "is_configured": true, 00:13:41.580 "data_offset": 2048, 00:13:41.580 "data_size": 63488 00:13:41.580 }, 00:13:41.580 { 00:13:41.580 "name": "BaseBdev2", 00:13:41.580 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:41.580 "is_configured": false, 00:13:41.580 "data_offset": 0, 00:13:41.580 "data_size": 0 00:13:41.580 }, 00:13:41.580 { 00:13:41.580 "name": "BaseBdev3", 00:13:41.580 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:41.580 "is_configured": false, 00:13:41.580 "data_offset": 0, 00:13:41.580 "data_size": 0 00:13:41.580 } 00:13:41.580 ] 00:13:41.580 }' 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:41.580 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:42.147 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:42.147 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:42.147 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:42.147 [2024-11-27 15:18:09.972365] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:42.147 [2024-11-27 15:18:09.972459] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:13:42.147 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:42.147 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:42.147 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:42.147 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:42.147 [2024-11-27 15:18:09.984373] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:42.147 [2024-11-27 15:18:09.986235] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:42.147 [2024-11-27 15:18:09.986315] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:42.147 [2024-11-27 15:18:09.986349] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:42.147 [2024-11-27 15:18:09.986376] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:42.147 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:42.147 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:13:42.147 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:42.147 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:42.147 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:42.147 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:42.147 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:42.147 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:42.147 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:42.147 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:42.147 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:42.147 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:42.147 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:42.147 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:42.147 15:18:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:42.147 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:42.147 15:18:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:42.147 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:42.147 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:42.147 "name": "Existed_Raid", 00:13:42.147 "uuid": "0df21cce-161a-4173-8542-64378dfcae50", 00:13:42.147 "strip_size_kb": 64, 00:13:42.147 "state": "configuring", 00:13:42.147 "raid_level": "raid5f", 00:13:42.147 "superblock": true, 00:13:42.147 "num_base_bdevs": 3, 00:13:42.147 "num_base_bdevs_discovered": 1, 00:13:42.147 "num_base_bdevs_operational": 3, 00:13:42.147 "base_bdevs_list": [ 00:13:42.147 { 00:13:42.147 "name": "BaseBdev1", 00:13:42.147 "uuid": "d97d609e-aef4-4a54-aaa2-2ff418116f8b", 00:13:42.147 "is_configured": true, 00:13:42.147 "data_offset": 2048, 00:13:42.147 "data_size": 63488 00:13:42.147 }, 00:13:42.147 { 00:13:42.147 "name": "BaseBdev2", 00:13:42.147 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:42.147 "is_configured": false, 00:13:42.147 "data_offset": 0, 00:13:42.147 "data_size": 0 00:13:42.147 }, 00:13:42.147 { 00:13:42.147 "name": "BaseBdev3", 00:13:42.147 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:42.147 "is_configured": false, 00:13:42.147 "data_offset": 0, 00:13:42.147 "data_size": 0 00:13:42.147 } 00:13:42.147 ] 00:13:42.147 }' 00:13:42.147 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:42.147 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:42.407 [2024-11-27 15:18:10.438741] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:42.407 BaseBdev2 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:42.407 [ 00:13:42.407 { 00:13:42.407 "name": "BaseBdev2", 00:13:42.407 "aliases": [ 00:13:42.407 "c392bce0-23ad-4663-9fae-93595494ab96" 00:13:42.407 ], 00:13:42.407 "product_name": "Malloc disk", 00:13:42.407 "block_size": 512, 00:13:42.407 "num_blocks": 65536, 00:13:42.407 "uuid": "c392bce0-23ad-4663-9fae-93595494ab96", 00:13:42.407 "assigned_rate_limits": { 00:13:42.407 "rw_ios_per_sec": 0, 00:13:42.407 "rw_mbytes_per_sec": 0, 00:13:42.407 "r_mbytes_per_sec": 0, 00:13:42.407 "w_mbytes_per_sec": 0 00:13:42.407 }, 00:13:42.407 "claimed": true, 00:13:42.407 "claim_type": "exclusive_write", 00:13:42.407 "zoned": false, 00:13:42.407 "supported_io_types": { 00:13:42.407 "read": true, 00:13:42.407 "write": true, 00:13:42.407 "unmap": true, 00:13:42.407 "flush": true, 00:13:42.407 "reset": true, 00:13:42.407 "nvme_admin": false, 00:13:42.407 "nvme_io": false, 00:13:42.407 "nvme_io_md": false, 00:13:42.407 "write_zeroes": true, 00:13:42.407 "zcopy": true, 00:13:42.407 "get_zone_info": false, 00:13:42.407 "zone_management": false, 00:13:42.407 "zone_append": false, 00:13:42.407 "compare": false, 00:13:42.407 "compare_and_write": false, 00:13:42.407 "abort": true, 00:13:42.407 "seek_hole": false, 00:13:42.407 "seek_data": false, 00:13:42.407 "copy": true, 00:13:42.407 "nvme_iov_md": false 00:13:42.407 }, 00:13:42.407 "memory_domains": [ 00:13:42.407 { 00:13:42.407 "dma_device_id": "system", 00:13:42.407 "dma_device_type": 1 00:13:42.407 }, 00:13:42.407 { 00:13:42.407 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:42.407 "dma_device_type": 2 00:13:42.407 } 00:13:42.407 ], 00:13:42.407 "driver_specific": {} 00:13:42.407 } 00:13:42.407 ] 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:42.407 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:42.667 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:42.667 "name": "Existed_Raid", 00:13:42.667 "uuid": "0df21cce-161a-4173-8542-64378dfcae50", 00:13:42.667 "strip_size_kb": 64, 00:13:42.667 "state": "configuring", 00:13:42.667 "raid_level": "raid5f", 00:13:42.667 "superblock": true, 00:13:42.667 "num_base_bdevs": 3, 00:13:42.667 "num_base_bdevs_discovered": 2, 00:13:42.667 "num_base_bdevs_operational": 3, 00:13:42.667 "base_bdevs_list": [ 00:13:42.667 { 00:13:42.667 "name": "BaseBdev1", 00:13:42.667 "uuid": "d97d609e-aef4-4a54-aaa2-2ff418116f8b", 00:13:42.667 "is_configured": true, 00:13:42.667 "data_offset": 2048, 00:13:42.667 "data_size": 63488 00:13:42.667 }, 00:13:42.667 { 00:13:42.667 "name": "BaseBdev2", 00:13:42.667 "uuid": "c392bce0-23ad-4663-9fae-93595494ab96", 00:13:42.667 "is_configured": true, 00:13:42.667 "data_offset": 2048, 00:13:42.667 "data_size": 63488 00:13:42.667 }, 00:13:42.667 { 00:13:42.667 "name": "BaseBdev3", 00:13:42.667 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:42.667 "is_configured": false, 00:13:42.667 "data_offset": 0, 00:13:42.667 "data_size": 0 00:13:42.667 } 00:13:42.667 ] 00:13:42.667 }' 00:13:42.667 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:42.667 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:42.926 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:13:42.926 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:42.926 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:42.926 [2024-11-27 15:18:10.954347] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:42.926 [2024-11-27 15:18:10.955186] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:13:42.926 BaseBdev3 00:13:42.926 [2024-11-27 15:18:10.955397] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:42.926 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:42.926 [2024-11-27 15:18:10.956550] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:13:42.926 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:13:42.926 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:13:42.926 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:42.926 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:13:42.926 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:42.926 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:42.926 [2024-11-27 15:18:10.958153] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:13:42.926 [2024-11-27 15:18:10.958201] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:13:42.926 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:42.926 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:42.926 [2024-11-27 15:18:10.958691] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:42.926 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:42.926 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:42.926 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:13:42.926 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:42.926 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:42.926 [ 00:13:42.926 { 00:13:42.926 "name": "BaseBdev3", 00:13:42.926 "aliases": [ 00:13:42.926 "ab0cf7dd-734e-4ce5-8ef3-7c9bc5b73497" 00:13:42.926 ], 00:13:42.926 "product_name": "Malloc disk", 00:13:42.926 "block_size": 512, 00:13:42.926 "num_blocks": 65536, 00:13:42.926 "uuid": "ab0cf7dd-734e-4ce5-8ef3-7c9bc5b73497", 00:13:42.926 "assigned_rate_limits": { 00:13:42.926 "rw_ios_per_sec": 0, 00:13:42.926 "rw_mbytes_per_sec": 0, 00:13:42.926 "r_mbytes_per_sec": 0, 00:13:42.926 "w_mbytes_per_sec": 0 00:13:42.926 }, 00:13:42.926 "claimed": true, 00:13:42.926 "claim_type": "exclusive_write", 00:13:42.926 "zoned": false, 00:13:42.926 "supported_io_types": { 00:13:42.926 "read": true, 00:13:42.926 "write": true, 00:13:42.926 "unmap": true, 00:13:42.926 "flush": true, 00:13:42.926 "reset": true, 00:13:42.926 "nvme_admin": false, 00:13:42.926 "nvme_io": false, 00:13:42.926 "nvme_io_md": false, 00:13:42.926 "write_zeroes": true, 00:13:42.926 "zcopy": true, 00:13:42.926 "get_zone_info": false, 00:13:42.926 "zone_management": false, 00:13:42.926 "zone_append": false, 00:13:42.926 "compare": false, 00:13:42.926 "compare_and_write": false, 00:13:42.926 "abort": true, 00:13:42.926 "seek_hole": false, 00:13:42.926 "seek_data": false, 00:13:42.926 "copy": true, 00:13:42.927 "nvme_iov_md": false 00:13:42.927 }, 00:13:42.927 "memory_domains": [ 00:13:42.927 { 00:13:42.927 "dma_device_id": "system", 00:13:42.927 "dma_device_type": 1 00:13:42.927 }, 00:13:42.927 { 00:13:42.927 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:42.927 "dma_device_type": 2 00:13:42.927 } 00:13:42.927 ], 00:13:42.927 "driver_specific": {} 00:13:42.927 } 00:13:42.927 ] 00:13:42.927 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:42.927 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:13:42.927 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:13:42.927 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:42.927 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:13:42.927 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:42.927 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:42.927 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:42.927 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:42.927 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:42.927 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:42.927 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:42.927 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:42.927 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:42.927 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:42.927 15:18:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:42.927 15:18:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:42.927 15:18:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:42.927 15:18:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.185 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:43.185 "name": "Existed_Raid", 00:13:43.185 "uuid": "0df21cce-161a-4173-8542-64378dfcae50", 00:13:43.185 "strip_size_kb": 64, 00:13:43.185 "state": "online", 00:13:43.185 "raid_level": "raid5f", 00:13:43.185 "superblock": true, 00:13:43.185 "num_base_bdevs": 3, 00:13:43.185 "num_base_bdevs_discovered": 3, 00:13:43.185 "num_base_bdevs_operational": 3, 00:13:43.185 "base_bdevs_list": [ 00:13:43.185 { 00:13:43.185 "name": "BaseBdev1", 00:13:43.185 "uuid": "d97d609e-aef4-4a54-aaa2-2ff418116f8b", 00:13:43.185 "is_configured": true, 00:13:43.185 "data_offset": 2048, 00:13:43.185 "data_size": 63488 00:13:43.185 }, 00:13:43.186 { 00:13:43.186 "name": "BaseBdev2", 00:13:43.186 "uuid": "c392bce0-23ad-4663-9fae-93595494ab96", 00:13:43.186 "is_configured": true, 00:13:43.186 "data_offset": 2048, 00:13:43.186 "data_size": 63488 00:13:43.186 }, 00:13:43.186 { 00:13:43.186 "name": "BaseBdev3", 00:13:43.186 "uuid": "ab0cf7dd-734e-4ce5-8ef3-7c9bc5b73497", 00:13:43.186 "is_configured": true, 00:13:43.186 "data_offset": 2048, 00:13:43.186 "data_size": 63488 00:13:43.186 } 00:13:43.186 ] 00:13:43.186 }' 00:13:43.186 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:43.186 15:18:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.445 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:13:43.445 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:13:43.445 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:43.445 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:43.445 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:13:43.445 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:43.445 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:13:43.445 15:18:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.445 15:18:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.445 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:43.445 [2024-11-27 15:18:11.429751] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:43.445 15:18:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.445 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:43.445 "name": "Existed_Raid", 00:13:43.445 "aliases": [ 00:13:43.445 "0df21cce-161a-4173-8542-64378dfcae50" 00:13:43.445 ], 00:13:43.445 "product_name": "Raid Volume", 00:13:43.445 "block_size": 512, 00:13:43.445 "num_blocks": 126976, 00:13:43.445 "uuid": "0df21cce-161a-4173-8542-64378dfcae50", 00:13:43.445 "assigned_rate_limits": { 00:13:43.445 "rw_ios_per_sec": 0, 00:13:43.445 "rw_mbytes_per_sec": 0, 00:13:43.445 "r_mbytes_per_sec": 0, 00:13:43.445 "w_mbytes_per_sec": 0 00:13:43.445 }, 00:13:43.445 "claimed": false, 00:13:43.445 "zoned": false, 00:13:43.445 "supported_io_types": { 00:13:43.445 "read": true, 00:13:43.445 "write": true, 00:13:43.445 "unmap": false, 00:13:43.445 "flush": false, 00:13:43.445 "reset": true, 00:13:43.445 "nvme_admin": false, 00:13:43.445 "nvme_io": false, 00:13:43.445 "nvme_io_md": false, 00:13:43.445 "write_zeroes": true, 00:13:43.445 "zcopy": false, 00:13:43.445 "get_zone_info": false, 00:13:43.445 "zone_management": false, 00:13:43.445 "zone_append": false, 00:13:43.445 "compare": false, 00:13:43.445 "compare_and_write": false, 00:13:43.445 "abort": false, 00:13:43.445 "seek_hole": false, 00:13:43.445 "seek_data": false, 00:13:43.445 "copy": false, 00:13:43.445 "nvme_iov_md": false 00:13:43.445 }, 00:13:43.445 "driver_specific": { 00:13:43.445 "raid": { 00:13:43.445 "uuid": "0df21cce-161a-4173-8542-64378dfcae50", 00:13:43.445 "strip_size_kb": 64, 00:13:43.445 "state": "online", 00:13:43.445 "raid_level": "raid5f", 00:13:43.445 "superblock": true, 00:13:43.445 "num_base_bdevs": 3, 00:13:43.445 "num_base_bdevs_discovered": 3, 00:13:43.445 "num_base_bdevs_operational": 3, 00:13:43.445 "base_bdevs_list": [ 00:13:43.445 { 00:13:43.445 "name": "BaseBdev1", 00:13:43.445 "uuid": "d97d609e-aef4-4a54-aaa2-2ff418116f8b", 00:13:43.445 "is_configured": true, 00:13:43.445 "data_offset": 2048, 00:13:43.445 "data_size": 63488 00:13:43.445 }, 00:13:43.445 { 00:13:43.445 "name": "BaseBdev2", 00:13:43.445 "uuid": "c392bce0-23ad-4663-9fae-93595494ab96", 00:13:43.445 "is_configured": true, 00:13:43.445 "data_offset": 2048, 00:13:43.445 "data_size": 63488 00:13:43.445 }, 00:13:43.445 { 00:13:43.445 "name": "BaseBdev3", 00:13:43.445 "uuid": "ab0cf7dd-734e-4ce5-8ef3-7c9bc5b73497", 00:13:43.445 "is_configured": true, 00:13:43.445 "data_offset": 2048, 00:13:43.445 "data_size": 63488 00:13:43.445 } 00:13:43.445 ] 00:13:43.445 } 00:13:43.445 } 00:13:43.445 }' 00:13:43.445 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:43.445 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:13:43.445 BaseBdev2 00:13:43.445 BaseBdev3' 00:13:43.445 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.710 [2024-11-27 15:18:11.713084] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 2 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:43.710 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:43.711 15:18:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.711 15:18:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.711 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:43.711 15:18:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.711 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:43.711 "name": "Existed_Raid", 00:13:43.711 "uuid": "0df21cce-161a-4173-8542-64378dfcae50", 00:13:43.711 "strip_size_kb": 64, 00:13:43.711 "state": "online", 00:13:43.711 "raid_level": "raid5f", 00:13:43.711 "superblock": true, 00:13:43.711 "num_base_bdevs": 3, 00:13:43.711 "num_base_bdevs_discovered": 2, 00:13:43.711 "num_base_bdevs_operational": 2, 00:13:43.711 "base_bdevs_list": [ 00:13:43.711 { 00:13:43.711 "name": null, 00:13:43.711 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:43.711 "is_configured": false, 00:13:43.711 "data_offset": 0, 00:13:43.711 "data_size": 63488 00:13:43.711 }, 00:13:43.711 { 00:13:43.711 "name": "BaseBdev2", 00:13:43.711 "uuid": "c392bce0-23ad-4663-9fae-93595494ab96", 00:13:43.711 "is_configured": true, 00:13:43.711 "data_offset": 2048, 00:13:43.711 "data_size": 63488 00:13:43.711 }, 00:13:43.711 { 00:13:43.711 "name": "BaseBdev3", 00:13:43.711 "uuid": "ab0cf7dd-734e-4ce5-8ef3-7c9bc5b73497", 00:13:43.711 "is_configured": true, 00:13:43.711 "data_offset": 2048, 00:13:43.711 "data_size": 63488 00:13:43.711 } 00:13:43.711 ] 00:13:43.711 }' 00:13:43.711 15:18:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:43.711 15:18:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:44.288 [2024-11-27 15:18:12.247666] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:44.288 [2024-11-27 15:18:12.247876] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:44.288 [2024-11-27 15:18:12.259218] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:44.288 [2024-11-27 15:18:12.299158] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:13:44.288 [2024-11-27 15:18:12.299241] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:44.288 BaseBdev2 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:44.288 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:44.548 [ 00:13:44.548 { 00:13:44.548 "name": "BaseBdev2", 00:13:44.548 "aliases": [ 00:13:44.548 "4d631b59-12dd-4a0f-b539-0440220c7811" 00:13:44.548 ], 00:13:44.548 "product_name": "Malloc disk", 00:13:44.548 "block_size": 512, 00:13:44.548 "num_blocks": 65536, 00:13:44.548 "uuid": "4d631b59-12dd-4a0f-b539-0440220c7811", 00:13:44.548 "assigned_rate_limits": { 00:13:44.548 "rw_ios_per_sec": 0, 00:13:44.548 "rw_mbytes_per_sec": 0, 00:13:44.548 "r_mbytes_per_sec": 0, 00:13:44.548 "w_mbytes_per_sec": 0 00:13:44.548 }, 00:13:44.548 "claimed": false, 00:13:44.548 "zoned": false, 00:13:44.548 "supported_io_types": { 00:13:44.548 "read": true, 00:13:44.548 "write": true, 00:13:44.548 "unmap": true, 00:13:44.548 "flush": true, 00:13:44.548 "reset": true, 00:13:44.548 "nvme_admin": false, 00:13:44.548 "nvme_io": false, 00:13:44.548 "nvme_io_md": false, 00:13:44.548 "write_zeroes": true, 00:13:44.548 "zcopy": true, 00:13:44.548 "get_zone_info": false, 00:13:44.548 "zone_management": false, 00:13:44.548 "zone_append": false, 00:13:44.548 "compare": false, 00:13:44.548 "compare_and_write": false, 00:13:44.548 "abort": true, 00:13:44.548 "seek_hole": false, 00:13:44.548 "seek_data": false, 00:13:44.548 "copy": true, 00:13:44.548 "nvme_iov_md": false 00:13:44.548 }, 00:13:44.548 "memory_domains": [ 00:13:44.548 { 00:13:44.548 "dma_device_id": "system", 00:13:44.548 "dma_device_type": 1 00:13:44.548 }, 00:13:44.548 { 00:13:44.548 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:44.548 "dma_device_type": 2 00:13:44.548 } 00:13:44.548 ], 00:13:44.548 "driver_specific": {} 00:13:44.548 } 00:13:44.548 ] 00:13:44.548 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:44.548 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:13:44.548 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:13:44.548 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:44.548 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:13:44.548 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:44.548 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:44.548 BaseBdev3 00:13:44.548 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:44.548 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:13:44.548 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:13:44.548 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:44.548 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:13:44.548 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:44.548 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:44.548 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:44.548 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:44.548 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:44.548 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:44.548 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:13:44.548 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:44.548 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:44.548 [ 00:13:44.548 { 00:13:44.548 "name": "BaseBdev3", 00:13:44.548 "aliases": [ 00:13:44.548 "e4efe88e-9f7c-4155-adef-81e9405d202f" 00:13:44.548 ], 00:13:44.548 "product_name": "Malloc disk", 00:13:44.548 "block_size": 512, 00:13:44.548 "num_blocks": 65536, 00:13:44.548 "uuid": "e4efe88e-9f7c-4155-adef-81e9405d202f", 00:13:44.548 "assigned_rate_limits": { 00:13:44.548 "rw_ios_per_sec": 0, 00:13:44.548 "rw_mbytes_per_sec": 0, 00:13:44.549 "r_mbytes_per_sec": 0, 00:13:44.549 "w_mbytes_per_sec": 0 00:13:44.549 }, 00:13:44.549 "claimed": false, 00:13:44.549 "zoned": false, 00:13:44.549 "supported_io_types": { 00:13:44.549 "read": true, 00:13:44.549 "write": true, 00:13:44.549 "unmap": true, 00:13:44.549 "flush": true, 00:13:44.549 "reset": true, 00:13:44.549 "nvme_admin": false, 00:13:44.549 "nvme_io": false, 00:13:44.549 "nvme_io_md": false, 00:13:44.549 "write_zeroes": true, 00:13:44.549 "zcopy": true, 00:13:44.549 "get_zone_info": false, 00:13:44.549 "zone_management": false, 00:13:44.549 "zone_append": false, 00:13:44.549 "compare": false, 00:13:44.549 "compare_and_write": false, 00:13:44.549 "abort": true, 00:13:44.549 "seek_hole": false, 00:13:44.549 "seek_data": false, 00:13:44.549 "copy": true, 00:13:44.549 "nvme_iov_md": false 00:13:44.549 }, 00:13:44.549 "memory_domains": [ 00:13:44.549 { 00:13:44.549 "dma_device_id": "system", 00:13:44.549 "dma_device_type": 1 00:13:44.549 }, 00:13:44.549 { 00:13:44.549 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:44.549 "dma_device_type": 2 00:13:44.549 } 00:13:44.549 ], 00:13:44.549 "driver_specific": {} 00:13:44.549 } 00:13:44.549 ] 00:13:44.549 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:44.549 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:13:44.549 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:13:44.549 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:44.549 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:44.549 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:44.549 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:44.549 [2024-11-27 15:18:12.470659] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:44.549 [2024-11-27 15:18:12.470745] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:44.549 [2024-11-27 15:18:12.470785] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:44.549 [2024-11-27 15:18:12.472604] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:44.549 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:44.549 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:44.549 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:44.549 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:44.549 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:44.549 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:44.549 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:44.549 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:44.549 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:44.549 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:44.549 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:44.549 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:44.549 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:44.549 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:44.549 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:44.549 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:44.549 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:44.549 "name": "Existed_Raid", 00:13:44.549 "uuid": "fcf2fcd9-f056-4eb0-9db9-bdc4504ef444", 00:13:44.549 "strip_size_kb": 64, 00:13:44.549 "state": "configuring", 00:13:44.549 "raid_level": "raid5f", 00:13:44.549 "superblock": true, 00:13:44.549 "num_base_bdevs": 3, 00:13:44.549 "num_base_bdevs_discovered": 2, 00:13:44.549 "num_base_bdevs_operational": 3, 00:13:44.549 "base_bdevs_list": [ 00:13:44.549 { 00:13:44.549 "name": "BaseBdev1", 00:13:44.549 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:44.549 "is_configured": false, 00:13:44.549 "data_offset": 0, 00:13:44.549 "data_size": 0 00:13:44.549 }, 00:13:44.549 { 00:13:44.549 "name": "BaseBdev2", 00:13:44.549 "uuid": "4d631b59-12dd-4a0f-b539-0440220c7811", 00:13:44.549 "is_configured": true, 00:13:44.549 "data_offset": 2048, 00:13:44.549 "data_size": 63488 00:13:44.549 }, 00:13:44.549 { 00:13:44.549 "name": "BaseBdev3", 00:13:44.549 "uuid": "e4efe88e-9f7c-4155-adef-81e9405d202f", 00:13:44.549 "is_configured": true, 00:13:44.549 "data_offset": 2048, 00:13:44.549 "data_size": 63488 00:13:44.549 } 00:13:44.549 ] 00:13:44.549 }' 00:13:44.549 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:44.549 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:45.119 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:13:45.119 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:45.119 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:45.119 [2024-11-27 15:18:12.973826] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:45.119 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:45.119 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:45.119 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:45.119 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:45.119 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:45.119 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:45.119 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:45.119 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:45.119 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:45.119 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:45.119 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:45.119 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:45.119 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:45.119 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:45.119 15:18:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:45.119 15:18:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:45.119 15:18:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:45.119 "name": "Existed_Raid", 00:13:45.119 "uuid": "fcf2fcd9-f056-4eb0-9db9-bdc4504ef444", 00:13:45.119 "strip_size_kb": 64, 00:13:45.119 "state": "configuring", 00:13:45.119 "raid_level": "raid5f", 00:13:45.119 "superblock": true, 00:13:45.119 "num_base_bdevs": 3, 00:13:45.119 "num_base_bdevs_discovered": 1, 00:13:45.119 "num_base_bdevs_operational": 3, 00:13:45.119 "base_bdevs_list": [ 00:13:45.119 { 00:13:45.119 "name": "BaseBdev1", 00:13:45.119 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:45.119 "is_configured": false, 00:13:45.119 "data_offset": 0, 00:13:45.119 "data_size": 0 00:13:45.119 }, 00:13:45.119 { 00:13:45.119 "name": null, 00:13:45.119 "uuid": "4d631b59-12dd-4a0f-b539-0440220c7811", 00:13:45.119 "is_configured": false, 00:13:45.119 "data_offset": 0, 00:13:45.119 "data_size": 63488 00:13:45.119 }, 00:13:45.119 { 00:13:45.119 "name": "BaseBdev3", 00:13:45.119 "uuid": "e4efe88e-9f7c-4155-adef-81e9405d202f", 00:13:45.119 "is_configured": true, 00:13:45.119 "data_offset": 2048, 00:13:45.119 "data_size": 63488 00:13:45.119 } 00:13:45.119 ] 00:13:45.119 }' 00:13:45.119 15:18:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:45.119 15:18:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:45.379 15:18:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:45.379 15:18:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:45.379 15:18:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:45.379 15:18:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:13:45.379 15:18:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:45.638 15:18:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:13:45.638 15:18:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:13:45.638 15:18:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:45.638 15:18:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:45.638 [2024-11-27 15:18:13.516222] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:45.638 BaseBdev1 00:13:45.638 15:18:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:45.638 15:18:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:13:45.638 15:18:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:13:45.638 15:18:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:45.638 15:18:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:13:45.638 15:18:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:45.638 15:18:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:45.638 15:18:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:45.638 15:18:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:45.638 15:18:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:45.638 15:18:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:45.638 15:18:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:45.638 15:18:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:45.638 15:18:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:45.638 [ 00:13:45.638 { 00:13:45.638 "name": "BaseBdev1", 00:13:45.638 "aliases": [ 00:13:45.638 "ae3f0714-0e8a-4500-aa0b-57f4b21ff7fd" 00:13:45.638 ], 00:13:45.639 "product_name": "Malloc disk", 00:13:45.639 "block_size": 512, 00:13:45.639 "num_blocks": 65536, 00:13:45.639 "uuid": "ae3f0714-0e8a-4500-aa0b-57f4b21ff7fd", 00:13:45.639 "assigned_rate_limits": { 00:13:45.639 "rw_ios_per_sec": 0, 00:13:45.639 "rw_mbytes_per_sec": 0, 00:13:45.639 "r_mbytes_per_sec": 0, 00:13:45.639 "w_mbytes_per_sec": 0 00:13:45.639 }, 00:13:45.639 "claimed": true, 00:13:45.639 "claim_type": "exclusive_write", 00:13:45.639 "zoned": false, 00:13:45.639 "supported_io_types": { 00:13:45.639 "read": true, 00:13:45.639 "write": true, 00:13:45.639 "unmap": true, 00:13:45.639 "flush": true, 00:13:45.639 "reset": true, 00:13:45.639 "nvme_admin": false, 00:13:45.639 "nvme_io": false, 00:13:45.639 "nvme_io_md": false, 00:13:45.639 "write_zeroes": true, 00:13:45.639 "zcopy": true, 00:13:45.639 "get_zone_info": false, 00:13:45.639 "zone_management": false, 00:13:45.639 "zone_append": false, 00:13:45.639 "compare": false, 00:13:45.639 "compare_and_write": false, 00:13:45.639 "abort": true, 00:13:45.639 "seek_hole": false, 00:13:45.639 "seek_data": false, 00:13:45.639 "copy": true, 00:13:45.639 "nvme_iov_md": false 00:13:45.639 }, 00:13:45.639 "memory_domains": [ 00:13:45.639 { 00:13:45.639 "dma_device_id": "system", 00:13:45.639 "dma_device_type": 1 00:13:45.639 }, 00:13:45.639 { 00:13:45.639 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:45.639 "dma_device_type": 2 00:13:45.639 } 00:13:45.639 ], 00:13:45.639 "driver_specific": {} 00:13:45.639 } 00:13:45.639 ] 00:13:45.639 15:18:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:45.639 15:18:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:13:45.639 15:18:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:45.639 15:18:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:45.639 15:18:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:45.639 15:18:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:45.639 15:18:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:45.639 15:18:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:45.639 15:18:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:45.639 15:18:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:45.639 15:18:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:45.639 15:18:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:45.639 15:18:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:45.639 15:18:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:45.639 15:18:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:45.639 15:18:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:45.639 15:18:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:45.639 15:18:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:45.639 "name": "Existed_Raid", 00:13:45.639 "uuid": "fcf2fcd9-f056-4eb0-9db9-bdc4504ef444", 00:13:45.639 "strip_size_kb": 64, 00:13:45.639 "state": "configuring", 00:13:45.639 "raid_level": "raid5f", 00:13:45.639 "superblock": true, 00:13:45.639 "num_base_bdevs": 3, 00:13:45.639 "num_base_bdevs_discovered": 2, 00:13:45.639 "num_base_bdevs_operational": 3, 00:13:45.639 "base_bdevs_list": [ 00:13:45.639 { 00:13:45.639 "name": "BaseBdev1", 00:13:45.639 "uuid": "ae3f0714-0e8a-4500-aa0b-57f4b21ff7fd", 00:13:45.639 "is_configured": true, 00:13:45.639 "data_offset": 2048, 00:13:45.639 "data_size": 63488 00:13:45.639 }, 00:13:45.639 { 00:13:45.639 "name": null, 00:13:45.639 "uuid": "4d631b59-12dd-4a0f-b539-0440220c7811", 00:13:45.639 "is_configured": false, 00:13:45.639 "data_offset": 0, 00:13:45.639 "data_size": 63488 00:13:45.639 }, 00:13:45.639 { 00:13:45.639 "name": "BaseBdev3", 00:13:45.639 "uuid": "e4efe88e-9f7c-4155-adef-81e9405d202f", 00:13:45.639 "is_configured": true, 00:13:45.639 "data_offset": 2048, 00:13:45.639 "data_size": 63488 00:13:45.639 } 00:13:45.639 ] 00:13:45.639 }' 00:13:45.639 15:18:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:45.639 15:18:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:45.898 15:18:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:45.899 15:18:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:13:45.899 15:18:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:45.899 15:18:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:45.899 15:18:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:45.899 15:18:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:13:45.899 15:18:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:13:45.899 15:18:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:45.899 15:18:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:45.899 [2024-11-27 15:18:13.999543] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:13:46.158 15:18:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:46.158 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:46.158 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:46.158 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:46.158 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:46.158 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:46.158 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:46.158 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:46.158 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:46.158 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:46.158 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:46.158 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:46.158 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:46.158 15:18:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:46.158 15:18:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.158 15:18:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:46.158 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:46.158 "name": "Existed_Raid", 00:13:46.158 "uuid": "fcf2fcd9-f056-4eb0-9db9-bdc4504ef444", 00:13:46.158 "strip_size_kb": 64, 00:13:46.158 "state": "configuring", 00:13:46.158 "raid_level": "raid5f", 00:13:46.158 "superblock": true, 00:13:46.158 "num_base_bdevs": 3, 00:13:46.158 "num_base_bdevs_discovered": 1, 00:13:46.158 "num_base_bdevs_operational": 3, 00:13:46.158 "base_bdevs_list": [ 00:13:46.158 { 00:13:46.158 "name": "BaseBdev1", 00:13:46.158 "uuid": "ae3f0714-0e8a-4500-aa0b-57f4b21ff7fd", 00:13:46.158 "is_configured": true, 00:13:46.158 "data_offset": 2048, 00:13:46.158 "data_size": 63488 00:13:46.158 }, 00:13:46.158 { 00:13:46.158 "name": null, 00:13:46.158 "uuid": "4d631b59-12dd-4a0f-b539-0440220c7811", 00:13:46.158 "is_configured": false, 00:13:46.158 "data_offset": 0, 00:13:46.158 "data_size": 63488 00:13:46.159 }, 00:13:46.159 { 00:13:46.159 "name": null, 00:13:46.159 "uuid": "e4efe88e-9f7c-4155-adef-81e9405d202f", 00:13:46.159 "is_configured": false, 00:13:46.159 "data_offset": 0, 00:13:46.159 "data_size": 63488 00:13:46.159 } 00:13:46.159 ] 00:13:46.159 }' 00:13:46.159 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:46.159 15:18:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.418 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:46.418 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:13:46.418 15:18:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:46.418 15:18:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.418 15:18:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:46.678 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:13:46.678 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:13:46.678 15:18:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:46.678 15:18:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.678 [2024-11-27 15:18:14.530925] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:46.678 15:18:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:46.678 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:46.678 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:46.678 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:46.678 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:46.678 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:46.678 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:46.678 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:46.678 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:46.678 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:46.678 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:46.678 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:46.678 15:18:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:46.678 15:18:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.678 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:46.678 15:18:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:46.678 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:46.678 "name": "Existed_Raid", 00:13:46.678 "uuid": "fcf2fcd9-f056-4eb0-9db9-bdc4504ef444", 00:13:46.678 "strip_size_kb": 64, 00:13:46.678 "state": "configuring", 00:13:46.678 "raid_level": "raid5f", 00:13:46.678 "superblock": true, 00:13:46.678 "num_base_bdevs": 3, 00:13:46.678 "num_base_bdevs_discovered": 2, 00:13:46.678 "num_base_bdevs_operational": 3, 00:13:46.678 "base_bdevs_list": [ 00:13:46.678 { 00:13:46.678 "name": "BaseBdev1", 00:13:46.678 "uuid": "ae3f0714-0e8a-4500-aa0b-57f4b21ff7fd", 00:13:46.678 "is_configured": true, 00:13:46.678 "data_offset": 2048, 00:13:46.678 "data_size": 63488 00:13:46.678 }, 00:13:46.678 { 00:13:46.678 "name": null, 00:13:46.678 "uuid": "4d631b59-12dd-4a0f-b539-0440220c7811", 00:13:46.678 "is_configured": false, 00:13:46.678 "data_offset": 0, 00:13:46.678 "data_size": 63488 00:13:46.678 }, 00:13:46.678 { 00:13:46.678 "name": "BaseBdev3", 00:13:46.678 "uuid": "e4efe88e-9f7c-4155-adef-81e9405d202f", 00:13:46.678 "is_configured": true, 00:13:46.678 "data_offset": 2048, 00:13:46.678 "data_size": 63488 00:13:46.678 } 00:13:46.678 ] 00:13:46.678 }' 00:13:46.678 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:46.678 15:18:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.938 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:46.938 15:18:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:46.938 15:18:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:13:46.938 15:18:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.938 15:18:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:46.938 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:13:46.938 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:13:46.938 15:18:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:46.938 15:18:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.938 [2024-11-27 15:18:15.026103] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:46.938 15:18:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:46.938 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:46.938 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:46.938 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:46.938 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:46.939 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:46.939 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:46.939 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:46.939 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:46.939 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:46.939 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:47.198 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:47.198 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:47.198 15:18:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.198 15:18:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:47.198 15:18:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:47.198 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:47.198 "name": "Existed_Raid", 00:13:47.198 "uuid": "fcf2fcd9-f056-4eb0-9db9-bdc4504ef444", 00:13:47.198 "strip_size_kb": 64, 00:13:47.198 "state": "configuring", 00:13:47.198 "raid_level": "raid5f", 00:13:47.198 "superblock": true, 00:13:47.198 "num_base_bdevs": 3, 00:13:47.198 "num_base_bdevs_discovered": 1, 00:13:47.198 "num_base_bdevs_operational": 3, 00:13:47.198 "base_bdevs_list": [ 00:13:47.198 { 00:13:47.198 "name": null, 00:13:47.198 "uuid": "ae3f0714-0e8a-4500-aa0b-57f4b21ff7fd", 00:13:47.198 "is_configured": false, 00:13:47.198 "data_offset": 0, 00:13:47.198 "data_size": 63488 00:13:47.198 }, 00:13:47.198 { 00:13:47.198 "name": null, 00:13:47.198 "uuid": "4d631b59-12dd-4a0f-b539-0440220c7811", 00:13:47.198 "is_configured": false, 00:13:47.198 "data_offset": 0, 00:13:47.198 "data_size": 63488 00:13:47.198 }, 00:13:47.198 { 00:13:47.198 "name": "BaseBdev3", 00:13:47.199 "uuid": "e4efe88e-9f7c-4155-adef-81e9405d202f", 00:13:47.199 "is_configured": true, 00:13:47.199 "data_offset": 2048, 00:13:47.199 "data_size": 63488 00:13:47.199 } 00:13:47.199 ] 00:13:47.199 }' 00:13:47.199 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:47.199 15:18:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:47.458 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:13:47.458 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:47.458 15:18:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.458 15:18:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:47.458 15:18:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:47.458 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:13:47.458 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:13:47.458 15:18:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.458 15:18:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:47.458 [2024-11-27 15:18:15.523871] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:47.458 15:18:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:47.458 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:47.458 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:47.458 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:47.458 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:47.458 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:47.458 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:47.458 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:47.458 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:47.458 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:47.458 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:47.458 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:47.458 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:47.458 15:18:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.458 15:18:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:47.458 15:18:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:47.718 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:47.718 "name": "Existed_Raid", 00:13:47.718 "uuid": "fcf2fcd9-f056-4eb0-9db9-bdc4504ef444", 00:13:47.718 "strip_size_kb": 64, 00:13:47.718 "state": "configuring", 00:13:47.718 "raid_level": "raid5f", 00:13:47.718 "superblock": true, 00:13:47.718 "num_base_bdevs": 3, 00:13:47.718 "num_base_bdevs_discovered": 2, 00:13:47.718 "num_base_bdevs_operational": 3, 00:13:47.718 "base_bdevs_list": [ 00:13:47.718 { 00:13:47.718 "name": null, 00:13:47.718 "uuid": "ae3f0714-0e8a-4500-aa0b-57f4b21ff7fd", 00:13:47.718 "is_configured": false, 00:13:47.718 "data_offset": 0, 00:13:47.718 "data_size": 63488 00:13:47.718 }, 00:13:47.718 { 00:13:47.718 "name": "BaseBdev2", 00:13:47.718 "uuid": "4d631b59-12dd-4a0f-b539-0440220c7811", 00:13:47.718 "is_configured": true, 00:13:47.718 "data_offset": 2048, 00:13:47.718 "data_size": 63488 00:13:47.718 }, 00:13:47.718 { 00:13:47.718 "name": "BaseBdev3", 00:13:47.718 "uuid": "e4efe88e-9f7c-4155-adef-81e9405d202f", 00:13:47.718 "is_configured": true, 00:13:47.718 "data_offset": 2048, 00:13:47.718 "data_size": 63488 00:13:47.718 } 00:13:47.718 ] 00:13:47.718 }' 00:13:47.718 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:47.718 15:18:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:47.977 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:47.977 15:18:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.978 15:18:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:47.978 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:13:47.978 15:18:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:47.978 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:13:47.978 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:13:47.978 15:18:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:47.978 15:18:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.978 15:18:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:47.978 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:47.978 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u ae3f0714-0e8a-4500-aa0b-57f4b21ff7fd 00:13:47.978 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.978 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:47.978 [2024-11-27 15:18:16.050190] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:13:47.978 [2024-11-27 15:18:16.050438] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:13:47.978 [2024-11-27 15:18:16.050494] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:47.978 [2024-11-27 15:18:16.050772] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:13:47.978 NewBaseBdev 00:13:47.978 [2024-11-27 15:18:16.051260] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:13:47.978 [2024-11-27 15:18:16.051311] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006d00 00:13:47.978 [2024-11-27 15:18:16.051480] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:47.978 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:47.978 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:13:47.978 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:13:47.978 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:47.978 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:13:47.978 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:47.978 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:47.978 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:47.978 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.978 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:47.978 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:47.978 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:13:47.978 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.978 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:47.978 [ 00:13:47.978 { 00:13:47.978 "name": "NewBaseBdev", 00:13:47.978 "aliases": [ 00:13:47.978 "ae3f0714-0e8a-4500-aa0b-57f4b21ff7fd" 00:13:47.978 ], 00:13:47.978 "product_name": "Malloc disk", 00:13:47.978 "block_size": 512, 00:13:47.978 "num_blocks": 65536, 00:13:47.978 "uuid": "ae3f0714-0e8a-4500-aa0b-57f4b21ff7fd", 00:13:47.978 "assigned_rate_limits": { 00:13:47.978 "rw_ios_per_sec": 0, 00:13:47.978 "rw_mbytes_per_sec": 0, 00:13:47.978 "r_mbytes_per_sec": 0, 00:13:47.978 "w_mbytes_per_sec": 0 00:13:47.978 }, 00:13:47.978 "claimed": true, 00:13:47.978 "claim_type": "exclusive_write", 00:13:47.978 "zoned": false, 00:13:47.978 "supported_io_types": { 00:13:47.978 "read": true, 00:13:47.978 "write": true, 00:13:47.978 "unmap": true, 00:13:47.978 "flush": true, 00:13:47.978 "reset": true, 00:13:47.978 "nvme_admin": false, 00:13:47.978 "nvme_io": false, 00:13:47.978 "nvme_io_md": false, 00:13:47.978 "write_zeroes": true, 00:13:48.237 "zcopy": true, 00:13:48.237 "get_zone_info": false, 00:13:48.237 "zone_management": false, 00:13:48.237 "zone_append": false, 00:13:48.237 "compare": false, 00:13:48.237 "compare_and_write": false, 00:13:48.237 "abort": true, 00:13:48.237 "seek_hole": false, 00:13:48.237 "seek_data": false, 00:13:48.237 "copy": true, 00:13:48.237 "nvme_iov_md": false 00:13:48.237 }, 00:13:48.238 "memory_domains": [ 00:13:48.238 { 00:13:48.238 "dma_device_id": "system", 00:13:48.238 "dma_device_type": 1 00:13:48.238 }, 00:13:48.238 { 00:13:48.238 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:48.238 "dma_device_type": 2 00:13:48.238 } 00:13:48.238 ], 00:13:48.238 "driver_specific": {} 00:13:48.238 } 00:13:48.238 ] 00:13:48.238 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:48.238 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:13:48.238 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:13:48.238 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:48.238 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:48.238 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:48.238 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:48.238 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:48.238 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:48.238 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:48.238 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:48.238 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:48.238 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:48.238 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:48.238 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:48.238 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:48.238 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:48.238 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:48.238 "name": "Existed_Raid", 00:13:48.238 "uuid": "fcf2fcd9-f056-4eb0-9db9-bdc4504ef444", 00:13:48.238 "strip_size_kb": 64, 00:13:48.238 "state": "online", 00:13:48.238 "raid_level": "raid5f", 00:13:48.238 "superblock": true, 00:13:48.238 "num_base_bdevs": 3, 00:13:48.238 "num_base_bdevs_discovered": 3, 00:13:48.238 "num_base_bdevs_operational": 3, 00:13:48.238 "base_bdevs_list": [ 00:13:48.238 { 00:13:48.238 "name": "NewBaseBdev", 00:13:48.238 "uuid": "ae3f0714-0e8a-4500-aa0b-57f4b21ff7fd", 00:13:48.238 "is_configured": true, 00:13:48.238 "data_offset": 2048, 00:13:48.238 "data_size": 63488 00:13:48.238 }, 00:13:48.238 { 00:13:48.238 "name": "BaseBdev2", 00:13:48.238 "uuid": "4d631b59-12dd-4a0f-b539-0440220c7811", 00:13:48.238 "is_configured": true, 00:13:48.238 "data_offset": 2048, 00:13:48.238 "data_size": 63488 00:13:48.238 }, 00:13:48.238 { 00:13:48.238 "name": "BaseBdev3", 00:13:48.238 "uuid": "e4efe88e-9f7c-4155-adef-81e9405d202f", 00:13:48.238 "is_configured": true, 00:13:48.238 "data_offset": 2048, 00:13:48.238 "data_size": 63488 00:13:48.238 } 00:13:48.238 ] 00:13:48.238 }' 00:13:48.238 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:48.238 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:48.497 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:13:48.497 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:13:48.497 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:48.497 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:48.497 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:13:48.497 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:48.498 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:48.498 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:13:48.498 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:48.498 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:48.498 [2024-11-27 15:18:16.497682] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:48.498 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:48.498 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:48.498 "name": "Existed_Raid", 00:13:48.498 "aliases": [ 00:13:48.498 "fcf2fcd9-f056-4eb0-9db9-bdc4504ef444" 00:13:48.498 ], 00:13:48.498 "product_name": "Raid Volume", 00:13:48.498 "block_size": 512, 00:13:48.498 "num_blocks": 126976, 00:13:48.498 "uuid": "fcf2fcd9-f056-4eb0-9db9-bdc4504ef444", 00:13:48.498 "assigned_rate_limits": { 00:13:48.498 "rw_ios_per_sec": 0, 00:13:48.498 "rw_mbytes_per_sec": 0, 00:13:48.498 "r_mbytes_per_sec": 0, 00:13:48.498 "w_mbytes_per_sec": 0 00:13:48.498 }, 00:13:48.498 "claimed": false, 00:13:48.498 "zoned": false, 00:13:48.498 "supported_io_types": { 00:13:48.498 "read": true, 00:13:48.498 "write": true, 00:13:48.498 "unmap": false, 00:13:48.498 "flush": false, 00:13:48.498 "reset": true, 00:13:48.498 "nvme_admin": false, 00:13:48.498 "nvme_io": false, 00:13:48.498 "nvme_io_md": false, 00:13:48.498 "write_zeroes": true, 00:13:48.498 "zcopy": false, 00:13:48.498 "get_zone_info": false, 00:13:48.498 "zone_management": false, 00:13:48.498 "zone_append": false, 00:13:48.498 "compare": false, 00:13:48.498 "compare_and_write": false, 00:13:48.498 "abort": false, 00:13:48.498 "seek_hole": false, 00:13:48.498 "seek_data": false, 00:13:48.498 "copy": false, 00:13:48.498 "nvme_iov_md": false 00:13:48.498 }, 00:13:48.498 "driver_specific": { 00:13:48.498 "raid": { 00:13:48.498 "uuid": "fcf2fcd9-f056-4eb0-9db9-bdc4504ef444", 00:13:48.498 "strip_size_kb": 64, 00:13:48.498 "state": "online", 00:13:48.498 "raid_level": "raid5f", 00:13:48.498 "superblock": true, 00:13:48.498 "num_base_bdevs": 3, 00:13:48.498 "num_base_bdevs_discovered": 3, 00:13:48.498 "num_base_bdevs_operational": 3, 00:13:48.498 "base_bdevs_list": [ 00:13:48.498 { 00:13:48.498 "name": "NewBaseBdev", 00:13:48.498 "uuid": "ae3f0714-0e8a-4500-aa0b-57f4b21ff7fd", 00:13:48.498 "is_configured": true, 00:13:48.498 "data_offset": 2048, 00:13:48.498 "data_size": 63488 00:13:48.498 }, 00:13:48.498 { 00:13:48.498 "name": "BaseBdev2", 00:13:48.498 "uuid": "4d631b59-12dd-4a0f-b539-0440220c7811", 00:13:48.498 "is_configured": true, 00:13:48.498 "data_offset": 2048, 00:13:48.498 "data_size": 63488 00:13:48.498 }, 00:13:48.498 { 00:13:48.498 "name": "BaseBdev3", 00:13:48.498 "uuid": "e4efe88e-9f7c-4155-adef-81e9405d202f", 00:13:48.498 "is_configured": true, 00:13:48.498 "data_offset": 2048, 00:13:48.498 "data_size": 63488 00:13:48.498 } 00:13:48.498 ] 00:13:48.498 } 00:13:48.498 } 00:13:48.498 }' 00:13:48.498 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:48.498 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:13:48.498 BaseBdev2 00:13:48.498 BaseBdev3' 00:13:48.498 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:48.758 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:48.758 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:48.758 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:13:48.758 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:48.758 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:48.758 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:48.758 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:48.758 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:48.758 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:48.758 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:48.758 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:48.758 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:13:48.758 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:48.758 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:48.758 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:48.758 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:48.758 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:48.758 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:48.758 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:13:48.758 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:48.758 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:48.758 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:48.758 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:48.758 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:48.758 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:48.758 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:48.758 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:48.758 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:48.759 [2024-11-27 15:18:16.784995] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:48.759 [2024-11-27 15:18:16.785024] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:48.759 [2024-11-27 15:18:16.785102] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:48.759 [2024-11-27 15:18:16.785343] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:48.759 [2024-11-27 15:18:16.785380] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name Existed_Raid, state offline 00:13:48.759 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:48.759 15:18:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 91239 00:13:48.759 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 91239 ']' 00:13:48.759 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 91239 00:13:48.759 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:13:48.759 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:48.759 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 91239 00:13:48.759 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:48.759 killing process with pid 91239 00:13:48.759 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:48.759 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 91239' 00:13:48.759 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 91239 00:13:48.759 [2024-11-27 15:18:16.835133] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:48.759 15:18:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 91239 00:13:49.019 [2024-11-27 15:18:16.868085] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:49.019 15:18:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:13:49.019 00:13:49.019 real 0m9.046s 00:13:49.019 user 0m15.431s 00:13:49.019 sys 0m1.967s 00:13:49.019 15:18:17 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:49.019 15:18:17 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:49.019 ************************************ 00:13:49.019 END TEST raid5f_state_function_test_sb 00:13:49.019 ************************************ 00:13:49.279 15:18:17 bdev_raid -- bdev/bdev_raid.sh@988 -- # run_test raid5f_superblock_test raid_superblock_test raid5f 3 00:13:49.279 15:18:17 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:13:49.279 15:18:17 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:49.279 15:18:17 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:49.279 ************************************ 00:13:49.279 START TEST raid5f_superblock_test 00:13:49.279 ************************************ 00:13:49.279 15:18:17 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid5f 3 00:13:49.279 15:18:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid5f 00:13:49.279 15:18:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:13:49.279 15:18:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:13:49.279 15:18:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:13:49.279 15:18:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:13:49.279 15:18:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:13:49.279 15:18:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:13:49.279 15:18:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:13:49.279 15:18:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:13:49.279 15:18:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:13:49.279 15:18:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:13:49.279 15:18:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:13:49.279 15:18:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:13:49.279 15:18:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid5f '!=' raid1 ']' 00:13:49.279 15:18:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:13:49.279 15:18:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:13:49.279 15:18:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=91843 00:13:49.279 15:18:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:13:49.279 15:18:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 91843 00:13:49.279 15:18:17 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 91843 ']' 00:13:49.279 15:18:17 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:49.279 15:18:17 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:13:49.279 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:49.279 15:18:17 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:49.279 15:18:17 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:13:49.279 15:18:17 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:49.279 [2024-11-27 15:18:17.236136] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:13:49.279 [2024-11-27 15:18:17.236255] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid91843 ] 00:13:49.539 [2024-11-27 15:18:17.407480] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:49.539 [2024-11-27 15:18:17.434617] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:49.539 [2024-11-27 15:18:17.478595] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:49.539 [2024-11-27 15:18:17.478637] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.110 malloc1 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.110 [2024-11-27 15:18:18.074844] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:13:50.110 [2024-11-27 15:18:18.075004] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:50.110 [2024-11-27 15:18:18.075046] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:13:50.110 [2024-11-27 15:18:18.075091] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:50.110 [2024-11-27 15:18:18.077352] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:50.110 [2024-11-27 15:18:18.077457] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:13:50.110 pt1 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.110 malloc2 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.110 [2024-11-27 15:18:18.107593] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:13:50.110 [2024-11-27 15:18:18.107691] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:50.110 [2024-11-27 15:18:18.107734] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:13:50.110 [2024-11-27 15:18:18.107766] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:50.110 [2024-11-27 15:18:18.109955] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:50.110 [2024-11-27 15:18:18.110027] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:13:50.110 pt2 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:13:50.110 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.111 malloc3 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.111 [2024-11-27 15:18:18.140197] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:13:50.111 [2024-11-27 15:18:18.140247] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:50.111 [2024-11-27 15:18:18.140265] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:13:50.111 [2024-11-27 15:18:18.140276] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:50.111 [2024-11-27 15:18:18.142279] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:50.111 [2024-11-27 15:18:18.142318] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:13:50.111 pt3 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.111 [2024-11-27 15:18:18.152241] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:13:50.111 [2024-11-27 15:18:18.154173] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:50.111 [2024-11-27 15:18:18.154281] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:13:50.111 [2024-11-27 15:18:18.154470] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:13:50.111 [2024-11-27 15:18:18.154522] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:50.111 [2024-11-27 15:18:18.154818] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:13:50.111 [2024-11-27 15:18:18.155266] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:13:50.111 [2024-11-27 15:18:18.155318] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:13:50.111 [2024-11-27 15:18:18.155485] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:50.111 "name": "raid_bdev1", 00:13:50.111 "uuid": "092ef957-a012-458a-a0e9-88659d9c53f6", 00:13:50.111 "strip_size_kb": 64, 00:13:50.111 "state": "online", 00:13:50.111 "raid_level": "raid5f", 00:13:50.111 "superblock": true, 00:13:50.111 "num_base_bdevs": 3, 00:13:50.111 "num_base_bdevs_discovered": 3, 00:13:50.111 "num_base_bdevs_operational": 3, 00:13:50.111 "base_bdevs_list": [ 00:13:50.111 { 00:13:50.111 "name": "pt1", 00:13:50.111 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:50.111 "is_configured": true, 00:13:50.111 "data_offset": 2048, 00:13:50.111 "data_size": 63488 00:13:50.111 }, 00:13:50.111 { 00:13:50.111 "name": "pt2", 00:13:50.111 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:50.111 "is_configured": true, 00:13:50.111 "data_offset": 2048, 00:13:50.111 "data_size": 63488 00:13:50.111 }, 00:13:50.111 { 00:13:50.111 "name": "pt3", 00:13:50.111 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:50.111 "is_configured": true, 00:13:50.111 "data_offset": 2048, 00:13:50.111 "data_size": 63488 00:13:50.111 } 00:13:50.111 ] 00:13:50.111 }' 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:50.111 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.681 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:13:50.681 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:13:50.681 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:50.681 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:50.681 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:13:50.681 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:50.681 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:50.681 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.681 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.681 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:50.681 [2024-11-27 15:18:18.584294] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:50.681 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.681 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:50.681 "name": "raid_bdev1", 00:13:50.681 "aliases": [ 00:13:50.681 "092ef957-a012-458a-a0e9-88659d9c53f6" 00:13:50.681 ], 00:13:50.681 "product_name": "Raid Volume", 00:13:50.681 "block_size": 512, 00:13:50.681 "num_blocks": 126976, 00:13:50.681 "uuid": "092ef957-a012-458a-a0e9-88659d9c53f6", 00:13:50.681 "assigned_rate_limits": { 00:13:50.681 "rw_ios_per_sec": 0, 00:13:50.681 "rw_mbytes_per_sec": 0, 00:13:50.681 "r_mbytes_per_sec": 0, 00:13:50.681 "w_mbytes_per_sec": 0 00:13:50.681 }, 00:13:50.681 "claimed": false, 00:13:50.681 "zoned": false, 00:13:50.681 "supported_io_types": { 00:13:50.681 "read": true, 00:13:50.681 "write": true, 00:13:50.681 "unmap": false, 00:13:50.681 "flush": false, 00:13:50.681 "reset": true, 00:13:50.681 "nvme_admin": false, 00:13:50.681 "nvme_io": false, 00:13:50.681 "nvme_io_md": false, 00:13:50.681 "write_zeroes": true, 00:13:50.681 "zcopy": false, 00:13:50.681 "get_zone_info": false, 00:13:50.681 "zone_management": false, 00:13:50.681 "zone_append": false, 00:13:50.681 "compare": false, 00:13:50.681 "compare_and_write": false, 00:13:50.681 "abort": false, 00:13:50.681 "seek_hole": false, 00:13:50.681 "seek_data": false, 00:13:50.681 "copy": false, 00:13:50.681 "nvme_iov_md": false 00:13:50.681 }, 00:13:50.681 "driver_specific": { 00:13:50.681 "raid": { 00:13:50.681 "uuid": "092ef957-a012-458a-a0e9-88659d9c53f6", 00:13:50.681 "strip_size_kb": 64, 00:13:50.681 "state": "online", 00:13:50.681 "raid_level": "raid5f", 00:13:50.681 "superblock": true, 00:13:50.681 "num_base_bdevs": 3, 00:13:50.681 "num_base_bdevs_discovered": 3, 00:13:50.681 "num_base_bdevs_operational": 3, 00:13:50.681 "base_bdevs_list": [ 00:13:50.681 { 00:13:50.681 "name": "pt1", 00:13:50.681 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:50.681 "is_configured": true, 00:13:50.681 "data_offset": 2048, 00:13:50.681 "data_size": 63488 00:13:50.681 }, 00:13:50.681 { 00:13:50.681 "name": "pt2", 00:13:50.681 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:50.681 "is_configured": true, 00:13:50.681 "data_offset": 2048, 00:13:50.681 "data_size": 63488 00:13:50.681 }, 00:13:50.681 { 00:13:50.681 "name": "pt3", 00:13:50.681 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:50.681 "is_configured": true, 00:13:50.681 "data_offset": 2048, 00:13:50.681 "data_size": 63488 00:13:50.681 } 00:13:50.681 ] 00:13:50.681 } 00:13:50.681 } 00:13:50.681 }' 00:13:50.681 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:50.681 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:13:50.681 pt2 00:13:50.681 pt3' 00:13:50.681 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:50.681 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:50.681 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:50.681 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:13:50.681 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.681 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.681 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:50.681 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.681 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:50.681 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:50.681 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:50.681 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:13:50.681 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.681 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.681 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:50.681 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.941 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:50.941 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:50.941 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:50.941 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:13:50.941 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.941 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.941 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:50.941 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.941 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:50.941 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:50.941 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:50.941 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.941 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.941 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:13:50.941 [2024-11-27 15:18:18.863816] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:50.941 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.941 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=092ef957-a012-458a-a0e9-88659d9c53f6 00:13:50.941 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 092ef957-a012-458a-a0e9-88659d9c53f6 ']' 00:13:50.941 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:50.941 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.941 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.941 [2024-11-27 15:18:18.915547] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:50.941 [2024-11-27 15:18:18.915606] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:50.942 [2024-11-27 15:18:18.915716] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:50.942 [2024-11-27 15:18:18.915805] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:50.942 [2024-11-27 15:18:18.915881] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:13:50.942 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.942 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:50.942 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:13:50.942 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.942 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.942 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.942 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:13:50.942 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:13:50.942 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:13:50.942 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:13:50.942 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.942 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.942 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.942 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:13:50.942 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:13:50.942 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.942 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.942 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.942 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:13:50.942 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:13:50.942 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.942 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.942 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.942 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:13:50.942 15:18:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:13:50.942 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.942 15:18:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.942 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.942 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:13:50.942 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:13:50.942 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:13:50.942 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:13:50.942 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:13:51.205 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:13:51.205 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:13:51.205 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:13:51.205 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:13:51.205 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:51.205 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.205 [2024-11-27 15:18:19.055331] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:13:51.205 [2024-11-27 15:18:19.057299] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:13:51.205 [2024-11-27 15:18:19.057404] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:13:51.205 [2024-11-27 15:18:19.057472] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:13:51.205 [2024-11-27 15:18:19.057554] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:13:51.205 [2024-11-27 15:18:19.057648] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:13:51.205 [2024-11-27 15:18:19.057694] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:51.206 [2024-11-27 15:18:19.057726] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state configuring 00:13:51.206 request: 00:13:51.206 { 00:13:51.206 "name": "raid_bdev1", 00:13:51.206 "raid_level": "raid5f", 00:13:51.206 "base_bdevs": [ 00:13:51.206 "malloc1", 00:13:51.206 "malloc2", 00:13:51.206 "malloc3" 00:13:51.206 ], 00:13:51.206 "strip_size_kb": 64, 00:13:51.206 "superblock": false, 00:13:51.206 "method": "bdev_raid_create", 00:13:51.206 "req_id": 1 00:13:51.206 } 00:13:51.206 Got JSON-RPC error response 00:13:51.206 response: 00:13:51.206 { 00:13:51.206 "code": -17, 00:13:51.206 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:13:51.206 } 00:13:51.206 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:13:51.206 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:13:51.206 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:13:51.206 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:13:51.206 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:13:51.206 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:51.206 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:51.206 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.206 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:13:51.206 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:51.206 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:13:51.206 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:13:51.206 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:13:51.206 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:51.206 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.206 [2024-11-27 15:18:19.115219] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:13:51.206 [2024-11-27 15:18:19.115305] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:51.206 [2024-11-27 15:18:19.115336] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:13:51.207 [2024-11-27 15:18:19.115364] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:51.207 [2024-11-27 15:18:19.117534] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:51.207 [2024-11-27 15:18:19.117610] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:13:51.207 [2024-11-27 15:18:19.117692] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:13:51.207 [2024-11-27 15:18:19.117744] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:13:51.207 pt1 00:13:51.207 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:51.207 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:13:51.207 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:51.207 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:51.207 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:51.207 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:51.207 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:51.207 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:51.207 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:51.207 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:51.207 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:51.207 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:51.207 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:51.207 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:51.207 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.207 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:51.207 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:51.207 "name": "raid_bdev1", 00:13:51.207 "uuid": "092ef957-a012-458a-a0e9-88659d9c53f6", 00:13:51.208 "strip_size_kb": 64, 00:13:51.208 "state": "configuring", 00:13:51.208 "raid_level": "raid5f", 00:13:51.208 "superblock": true, 00:13:51.208 "num_base_bdevs": 3, 00:13:51.208 "num_base_bdevs_discovered": 1, 00:13:51.208 "num_base_bdevs_operational": 3, 00:13:51.208 "base_bdevs_list": [ 00:13:51.208 { 00:13:51.208 "name": "pt1", 00:13:51.208 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:51.208 "is_configured": true, 00:13:51.208 "data_offset": 2048, 00:13:51.208 "data_size": 63488 00:13:51.208 }, 00:13:51.208 { 00:13:51.208 "name": null, 00:13:51.208 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:51.208 "is_configured": false, 00:13:51.208 "data_offset": 2048, 00:13:51.208 "data_size": 63488 00:13:51.208 }, 00:13:51.208 { 00:13:51.208 "name": null, 00:13:51.208 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:51.208 "is_configured": false, 00:13:51.208 "data_offset": 2048, 00:13:51.208 "data_size": 63488 00:13:51.208 } 00:13:51.208 ] 00:13:51.208 }' 00:13:51.208 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:51.208 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.468 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:13:51.468 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:13:51.468 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:51.468 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.468 [2024-11-27 15:18:19.550507] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:13:51.468 [2024-11-27 15:18:19.550599] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:51.468 [2024-11-27 15:18:19.550633] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:13:51.468 [2024-11-27 15:18:19.550686] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:51.468 [2024-11-27 15:18:19.551084] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:51.468 [2024-11-27 15:18:19.551144] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:13:51.468 [2024-11-27 15:18:19.551232] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:13:51.468 [2024-11-27 15:18:19.551290] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:51.468 pt2 00:13:51.468 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:51.468 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:13:51.468 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:51.468 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.468 [2024-11-27 15:18:19.562477] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:13:51.468 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:51.468 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:13:51.468 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:51.468 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:51.468 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:51.468 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:51.468 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:51.468 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:51.468 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:51.468 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:51.468 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:51.728 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:51.728 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:51.728 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.728 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:51.728 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:51.728 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:51.728 "name": "raid_bdev1", 00:13:51.728 "uuid": "092ef957-a012-458a-a0e9-88659d9c53f6", 00:13:51.728 "strip_size_kb": 64, 00:13:51.728 "state": "configuring", 00:13:51.728 "raid_level": "raid5f", 00:13:51.728 "superblock": true, 00:13:51.728 "num_base_bdevs": 3, 00:13:51.728 "num_base_bdevs_discovered": 1, 00:13:51.728 "num_base_bdevs_operational": 3, 00:13:51.728 "base_bdevs_list": [ 00:13:51.728 { 00:13:51.728 "name": "pt1", 00:13:51.728 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:51.728 "is_configured": true, 00:13:51.728 "data_offset": 2048, 00:13:51.728 "data_size": 63488 00:13:51.728 }, 00:13:51.728 { 00:13:51.728 "name": null, 00:13:51.728 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:51.728 "is_configured": false, 00:13:51.728 "data_offset": 0, 00:13:51.728 "data_size": 63488 00:13:51.728 }, 00:13:51.728 { 00:13:51.728 "name": null, 00:13:51.728 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:51.728 "is_configured": false, 00:13:51.728 "data_offset": 2048, 00:13:51.728 "data_size": 63488 00:13:51.728 } 00:13:51.728 ] 00:13:51.728 }' 00:13:51.728 15:18:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:51.728 15:18:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.988 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:13:51.988 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:13:51.988 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:13:51.988 15:18:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:51.988 15:18:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.988 [2024-11-27 15:18:20.037694] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:13:51.988 [2024-11-27 15:18:20.037818] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:51.988 [2024-11-27 15:18:20.037857] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:13:51.988 [2024-11-27 15:18:20.037884] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:51.988 [2024-11-27 15:18:20.038337] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:51.988 [2024-11-27 15:18:20.038396] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:13:51.988 [2024-11-27 15:18:20.038502] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:13:51.988 [2024-11-27 15:18:20.038551] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:51.988 pt2 00:13:51.988 15:18:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:51.988 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:13:51.988 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:13:51.988 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:13:51.988 15:18:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:51.988 15:18:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.988 [2024-11-27 15:18:20.049635] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:13:51.988 [2024-11-27 15:18:20.049716] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:51.988 [2024-11-27 15:18:20.049749] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:13:51.988 [2024-11-27 15:18:20.049774] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:51.988 [2024-11-27 15:18:20.050132] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:51.988 [2024-11-27 15:18:20.050187] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:13:51.988 [2024-11-27 15:18:20.050268] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:13:51.988 [2024-11-27 15:18:20.050312] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:13:51.988 [2024-11-27 15:18:20.050446] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:13:51.988 [2024-11-27 15:18:20.050487] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:51.988 [2024-11-27 15:18:20.050723] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:13:51.988 [2024-11-27 15:18:20.051159] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:13:51.988 [2024-11-27 15:18:20.051178] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:13:51.988 [2024-11-27 15:18:20.051281] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:51.988 pt3 00:13:51.988 15:18:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:51.988 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:13:51.988 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:13:51.989 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:51.989 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:51.989 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:51.989 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:51.989 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:51.989 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:51.989 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:51.989 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:51.989 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:51.989 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:51.989 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:51.989 15:18:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:51.989 15:18:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.989 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:51.989 15:18:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.248 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:52.248 "name": "raid_bdev1", 00:13:52.248 "uuid": "092ef957-a012-458a-a0e9-88659d9c53f6", 00:13:52.248 "strip_size_kb": 64, 00:13:52.248 "state": "online", 00:13:52.248 "raid_level": "raid5f", 00:13:52.248 "superblock": true, 00:13:52.248 "num_base_bdevs": 3, 00:13:52.248 "num_base_bdevs_discovered": 3, 00:13:52.248 "num_base_bdevs_operational": 3, 00:13:52.248 "base_bdevs_list": [ 00:13:52.248 { 00:13:52.248 "name": "pt1", 00:13:52.248 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:52.248 "is_configured": true, 00:13:52.248 "data_offset": 2048, 00:13:52.248 "data_size": 63488 00:13:52.248 }, 00:13:52.248 { 00:13:52.248 "name": "pt2", 00:13:52.248 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:52.248 "is_configured": true, 00:13:52.248 "data_offset": 2048, 00:13:52.248 "data_size": 63488 00:13:52.248 }, 00:13:52.248 { 00:13:52.248 "name": "pt3", 00:13:52.248 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:52.248 "is_configured": true, 00:13:52.248 "data_offset": 2048, 00:13:52.248 "data_size": 63488 00:13:52.248 } 00:13:52.248 ] 00:13:52.248 }' 00:13:52.248 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:52.248 15:18:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.508 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:13:52.508 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:13:52.508 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:52.509 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:52.509 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:13:52.509 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:52.509 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:52.509 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:52.509 15:18:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.509 15:18:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.509 [2024-11-27 15:18:20.493130] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:52.509 15:18:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.509 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:52.509 "name": "raid_bdev1", 00:13:52.509 "aliases": [ 00:13:52.509 "092ef957-a012-458a-a0e9-88659d9c53f6" 00:13:52.509 ], 00:13:52.509 "product_name": "Raid Volume", 00:13:52.509 "block_size": 512, 00:13:52.509 "num_blocks": 126976, 00:13:52.509 "uuid": "092ef957-a012-458a-a0e9-88659d9c53f6", 00:13:52.509 "assigned_rate_limits": { 00:13:52.509 "rw_ios_per_sec": 0, 00:13:52.509 "rw_mbytes_per_sec": 0, 00:13:52.509 "r_mbytes_per_sec": 0, 00:13:52.509 "w_mbytes_per_sec": 0 00:13:52.509 }, 00:13:52.509 "claimed": false, 00:13:52.509 "zoned": false, 00:13:52.509 "supported_io_types": { 00:13:52.509 "read": true, 00:13:52.509 "write": true, 00:13:52.509 "unmap": false, 00:13:52.509 "flush": false, 00:13:52.509 "reset": true, 00:13:52.509 "nvme_admin": false, 00:13:52.509 "nvme_io": false, 00:13:52.509 "nvme_io_md": false, 00:13:52.509 "write_zeroes": true, 00:13:52.509 "zcopy": false, 00:13:52.509 "get_zone_info": false, 00:13:52.509 "zone_management": false, 00:13:52.509 "zone_append": false, 00:13:52.509 "compare": false, 00:13:52.509 "compare_and_write": false, 00:13:52.509 "abort": false, 00:13:52.509 "seek_hole": false, 00:13:52.509 "seek_data": false, 00:13:52.509 "copy": false, 00:13:52.509 "nvme_iov_md": false 00:13:52.509 }, 00:13:52.509 "driver_specific": { 00:13:52.509 "raid": { 00:13:52.509 "uuid": "092ef957-a012-458a-a0e9-88659d9c53f6", 00:13:52.509 "strip_size_kb": 64, 00:13:52.509 "state": "online", 00:13:52.509 "raid_level": "raid5f", 00:13:52.509 "superblock": true, 00:13:52.509 "num_base_bdevs": 3, 00:13:52.509 "num_base_bdevs_discovered": 3, 00:13:52.509 "num_base_bdevs_operational": 3, 00:13:52.509 "base_bdevs_list": [ 00:13:52.509 { 00:13:52.509 "name": "pt1", 00:13:52.509 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:52.509 "is_configured": true, 00:13:52.509 "data_offset": 2048, 00:13:52.509 "data_size": 63488 00:13:52.509 }, 00:13:52.509 { 00:13:52.509 "name": "pt2", 00:13:52.509 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:52.509 "is_configured": true, 00:13:52.509 "data_offset": 2048, 00:13:52.509 "data_size": 63488 00:13:52.509 }, 00:13:52.509 { 00:13:52.509 "name": "pt3", 00:13:52.509 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:52.509 "is_configured": true, 00:13:52.509 "data_offset": 2048, 00:13:52.509 "data_size": 63488 00:13:52.509 } 00:13:52.509 ] 00:13:52.509 } 00:13:52.509 } 00:13:52.509 }' 00:13:52.509 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:52.509 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:13:52.509 pt2 00:13:52.509 pt3' 00:13:52.509 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:13:52.770 [2024-11-27 15:18:20.756633] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 092ef957-a012-458a-a0e9-88659d9c53f6 '!=' 092ef957-a012-458a-a0e9-88659d9c53f6 ']' 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid5f 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.770 [2024-11-27 15:18:20.800425] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.770 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:52.770 "name": "raid_bdev1", 00:13:52.770 "uuid": "092ef957-a012-458a-a0e9-88659d9c53f6", 00:13:52.770 "strip_size_kb": 64, 00:13:52.770 "state": "online", 00:13:52.770 "raid_level": "raid5f", 00:13:52.770 "superblock": true, 00:13:52.770 "num_base_bdevs": 3, 00:13:52.770 "num_base_bdevs_discovered": 2, 00:13:52.770 "num_base_bdevs_operational": 2, 00:13:52.770 "base_bdevs_list": [ 00:13:52.770 { 00:13:52.770 "name": null, 00:13:52.770 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:52.770 "is_configured": false, 00:13:52.770 "data_offset": 0, 00:13:52.770 "data_size": 63488 00:13:52.770 }, 00:13:52.770 { 00:13:52.770 "name": "pt2", 00:13:52.770 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:52.770 "is_configured": true, 00:13:52.770 "data_offset": 2048, 00:13:52.770 "data_size": 63488 00:13:52.770 }, 00:13:52.770 { 00:13:52.770 "name": "pt3", 00:13:52.771 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:52.771 "is_configured": true, 00:13:52.771 "data_offset": 2048, 00:13:52.771 "data_size": 63488 00:13:52.771 } 00:13:52.771 ] 00:13:52.771 }' 00:13:52.771 15:18:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:52.771 15:18:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.341 [2024-11-27 15:18:21.271625] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:53.341 [2024-11-27 15:18:21.271698] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:53.341 [2024-11-27 15:18:21.271790] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:53.341 [2024-11-27 15:18:21.271865] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:53.341 [2024-11-27 15:18:21.272022] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.341 [2024-11-27 15:18:21.359521] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:13:53.341 [2024-11-27 15:18:21.359623] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:53.341 [2024-11-27 15:18:21.359658] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:13:53.341 [2024-11-27 15:18:21.359684] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:53.341 [2024-11-27 15:18:21.361778] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:53.341 [2024-11-27 15:18:21.361859] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:13:53.341 [2024-11-27 15:18:21.361979] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:13:53.341 [2024-11-27 15:18:21.362031] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:53.341 pt2 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 2 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:53.341 "name": "raid_bdev1", 00:13:53.341 "uuid": "092ef957-a012-458a-a0e9-88659d9c53f6", 00:13:53.341 "strip_size_kb": 64, 00:13:53.341 "state": "configuring", 00:13:53.341 "raid_level": "raid5f", 00:13:53.341 "superblock": true, 00:13:53.341 "num_base_bdevs": 3, 00:13:53.341 "num_base_bdevs_discovered": 1, 00:13:53.341 "num_base_bdevs_operational": 2, 00:13:53.341 "base_bdevs_list": [ 00:13:53.341 { 00:13:53.341 "name": null, 00:13:53.341 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:53.341 "is_configured": false, 00:13:53.341 "data_offset": 2048, 00:13:53.341 "data_size": 63488 00:13:53.341 }, 00:13:53.341 { 00:13:53.341 "name": "pt2", 00:13:53.341 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:53.341 "is_configured": true, 00:13:53.341 "data_offset": 2048, 00:13:53.341 "data_size": 63488 00:13:53.341 }, 00:13:53.341 { 00:13:53.341 "name": null, 00:13:53.341 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:53.341 "is_configured": false, 00:13:53.341 "data_offset": 2048, 00:13:53.341 "data_size": 63488 00:13:53.341 } 00:13:53.341 ] 00:13:53.341 }' 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:53.341 15:18:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.913 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:13:53.913 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:13:53.913 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@519 -- # i=2 00:13:53.913 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:13:53.913 15:18:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:53.913 15:18:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.913 [2024-11-27 15:18:21.826741] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:13:53.913 [2024-11-27 15:18:21.826869] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:53.913 [2024-11-27 15:18:21.826925] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:13:53.913 [2024-11-27 15:18:21.826968] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:53.913 [2024-11-27 15:18:21.827419] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:53.913 [2024-11-27 15:18:21.827488] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:13:53.913 [2024-11-27 15:18:21.827606] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:13:53.913 [2024-11-27 15:18:21.827659] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:13:53.913 [2024-11-27 15:18:21.827776] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:13:53.913 [2024-11-27 15:18:21.827812] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:53.913 [2024-11-27 15:18:21.828095] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:13:53.913 [2024-11-27 15:18:21.828589] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:13:53.913 [2024-11-27 15:18:21.828642] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006d00 00:13:53.913 [2024-11-27 15:18:21.828898] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:53.913 pt3 00:13:53.913 15:18:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:53.913 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:53.913 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:53.913 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:53.913 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:53.913 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:53.913 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:53.913 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:53.913 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:53.913 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:53.913 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:53.913 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:53.913 15:18:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:53.913 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:53.913 15:18:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.913 15:18:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:53.913 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:53.913 "name": "raid_bdev1", 00:13:53.913 "uuid": "092ef957-a012-458a-a0e9-88659d9c53f6", 00:13:53.913 "strip_size_kb": 64, 00:13:53.913 "state": "online", 00:13:53.913 "raid_level": "raid5f", 00:13:53.913 "superblock": true, 00:13:53.913 "num_base_bdevs": 3, 00:13:53.913 "num_base_bdevs_discovered": 2, 00:13:53.913 "num_base_bdevs_operational": 2, 00:13:53.913 "base_bdevs_list": [ 00:13:53.913 { 00:13:53.913 "name": null, 00:13:53.913 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:53.913 "is_configured": false, 00:13:53.913 "data_offset": 2048, 00:13:53.913 "data_size": 63488 00:13:53.913 }, 00:13:53.913 { 00:13:53.913 "name": "pt2", 00:13:53.913 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:53.913 "is_configured": true, 00:13:53.913 "data_offset": 2048, 00:13:53.913 "data_size": 63488 00:13:53.913 }, 00:13:53.913 { 00:13:53.913 "name": "pt3", 00:13:53.913 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:53.913 "is_configured": true, 00:13:53.913 "data_offset": 2048, 00:13:53.913 "data_size": 63488 00:13:53.913 } 00:13:53.913 ] 00:13:53.913 }' 00:13:53.913 15:18:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:53.913 15:18:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.204 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:54.205 15:18:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:54.205 15:18:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.205 [2024-11-27 15:18:22.257979] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:54.205 [2024-11-27 15:18:22.258045] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:54.205 [2024-11-27 15:18:22.258161] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:54.205 [2024-11-27 15:18:22.258248] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:54.205 [2024-11-27 15:18:22.258301] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name raid_bdev1, state offline 00:13:54.205 15:18:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:54.205 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:54.205 15:18:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:54.205 15:18:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.205 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:13:54.205 15:18:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:54.473 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:13:54.473 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:13:54.473 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 3 -gt 2 ']' 00:13:54.473 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@534 -- # i=2 00:13:54.473 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt3 00:13:54.473 15:18:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:54.473 15:18:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.473 15:18:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:54.473 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:13:54.473 15:18:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:54.473 15:18:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.473 [2024-11-27 15:18:22.329852] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:13:54.473 [2024-11-27 15:18:22.329984] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:54.473 [2024-11-27 15:18:22.330024] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:13:54.473 [2024-11-27 15:18:22.330084] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:54.473 [2024-11-27 15:18:22.332651] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:54.473 [2024-11-27 15:18:22.332741] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:13:54.473 [2024-11-27 15:18:22.332853] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:13:54.473 [2024-11-27 15:18:22.332966] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:13:54.473 [2024-11-27 15:18:22.333145] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:13:54.473 [2024-11-27 15:18:22.333220] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:54.473 [2024-11-27 15:18:22.333306] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007080 name raid_bdev1, state configuring 00:13:54.473 [2024-11-27 15:18:22.333407] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:54.473 pt1 00:13:54.473 15:18:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:54.473 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 3 -gt 2 ']' 00:13:54.473 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 2 00:13:54.473 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:54.474 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:54.474 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:54.474 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:54.474 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:54.474 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:54.474 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:54.474 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:54.474 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:54.474 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:54.474 15:18:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:54.474 15:18:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.474 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:54.474 15:18:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:54.474 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:54.474 "name": "raid_bdev1", 00:13:54.474 "uuid": "092ef957-a012-458a-a0e9-88659d9c53f6", 00:13:54.474 "strip_size_kb": 64, 00:13:54.474 "state": "configuring", 00:13:54.474 "raid_level": "raid5f", 00:13:54.474 "superblock": true, 00:13:54.474 "num_base_bdevs": 3, 00:13:54.474 "num_base_bdevs_discovered": 1, 00:13:54.474 "num_base_bdevs_operational": 2, 00:13:54.474 "base_bdevs_list": [ 00:13:54.474 { 00:13:54.474 "name": null, 00:13:54.474 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:54.474 "is_configured": false, 00:13:54.474 "data_offset": 2048, 00:13:54.474 "data_size": 63488 00:13:54.474 }, 00:13:54.474 { 00:13:54.474 "name": "pt2", 00:13:54.474 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:54.474 "is_configured": true, 00:13:54.474 "data_offset": 2048, 00:13:54.474 "data_size": 63488 00:13:54.474 }, 00:13:54.474 { 00:13:54.474 "name": null, 00:13:54.474 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:54.474 "is_configured": false, 00:13:54.474 "data_offset": 2048, 00:13:54.474 "data_size": 63488 00:13:54.474 } 00:13:54.474 ] 00:13:54.474 }' 00:13:54.474 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:54.474 15:18:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.733 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:13:54.733 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:13:54.733 15:18:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:54.733 15:18:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.733 15:18:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:54.733 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:13:54.733 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:13:54.733 15:18:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:54.733 15:18:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.993 [2024-11-27 15:18:22.841092] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:13:54.993 [2024-11-27 15:18:22.841294] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:54.993 [2024-11-27 15:18:22.841345] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:13:54.993 [2024-11-27 15:18:22.841392] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:54.993 [2024-11-27 15:18:22.842019] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:54.993 [2024-11-27 15:18:22.842103] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:13:54.993 [2024-11-27 15:18:22.842248] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:13:54.993 [2024-11-27 15:18:22.842316] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:13:54.993 [2024-11-27 15:18:22.842479] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007400 00:13:54.993 [2024-11-27 15:18:22.842527] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:54.993 [2024-11-27 15:18:22.842833] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:13:54.993 [2024-11-27 15:18:22.843398] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007400 00:13:54.993 [2024-11-27 15:18:22.843474] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007400 00:13:54.993 [2024-11-27 15:18:22.843737] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:54.993 pt3 00:13:54.993 15:18:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:54.993 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:54.993 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:54.993 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:54.993 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:54.993 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:54.993 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:54.993 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:54.993 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:54.993 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:54.993 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:54.993 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:54.993 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:54.993 15:18:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:54.993 15:18:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.993 15:18:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:54.993 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:54.993 "name": "raid_bdev1", 00:13:54.993 "uuid": "092ef957-a012-458a-a0e9-88659d9c53f6", 00:13:54.993 "strip_size_kb": 64, 00:13:54.993 "state": "online", 00:13:54.993 "raid_level": "raid5f", 00:13:54.993 "superblock": true, 00:13:54.993 "num_base_bdevs": 3, 00:13:54.993 "num_base_bdevs_discovered": 2, 00:13:54.993 "num_base_bdevs_operational": 2, 00:13:54.993 "base_bdevs_list": [ 00:13:54.993 { 00:13:54.993 "name": null, 00:13:54.993 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:54.993 "is_configured": false, 00:13:54.993 "data_offset": 2048, 00:13:54.993 "data_size": 63488 00:13:54.993 }, 00:13:54.993 { 00:13:54.993 "name": "pt2", 00:13:54.993 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:54.993 "is_configured": true, 00:13:54.993 "data_offset": 2048, 00:13:54.993 "data_size": 63488 00:13:54.993 }, 00:13:54.993 { 00:13:54.993 "name": "pt3", 00:13:54.993 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:54.993 "is_configured": true, 00:13:54.993 "data_offset": 2048, 00:13:54.993 "data_size": 63488 00:13:54.993 } 00:13:54.993 ] 00:13:54.993 }' 00:13:54.993 15:18:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:54.993 15:18:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:55.253 15:18:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:13:55.253 15:18:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:55.253 15:18:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:55.253 15:18:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:13:55.253 15:18:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:55.513 15:18:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:13:55.513 15:18:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:55.513 15:18:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:55.513 15:18:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:55.513 15:18:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:13:55.513 [2024-11-27 15:18:23.373342] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:55.513 15:18:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:55.513 15:18:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 092ef957-a012-458a-a0e9-88659d9c53f6 '!=' 092ef957-a012-458a-a0e9-88659d9c53f6 ']' 00:13:55.513 15:18:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 91843 00:13:55.513 15:18:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 91843 ']' 00:13:55.513 15:18:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@958 -- # kill -0 91843 00:13:55.513 15:18:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@959 -- # uname 00:13:55.513 15:18:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:55.513 15:18:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 91843 00:13:55.513 killing process with pid 91843 00:13:55.513 15:18:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:55.513 15:18:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:55.513 15:18:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 91843' 00:13:55.513 15:18:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@973 -- # kill 91843 00:13:55.513 [2024-11-27 15:18:23.450171] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:55.513 [2024-11-27 15:18:23.450308] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:55.514 15:18:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@978 -- # wait 91843 00:13:55.514 [2024-11-27 15:18:23.450393] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:55.514 [2024-11-27 15:18:23.450405] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007400 name raid_bdev1, state offline 00:13:55.514 [2024-11-27 15:18:23.513586] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:55.774 15:18:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:13:55.774 00:13:55.774 real 0m6.709s 00:13:55.774 user 0m11.067s 00:13:55.774 sys 0m1.420s 00:13:55.774 15:18:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:55.774 15:18:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:55.774 ************************************ 00:13:55.774 END TEST raid5f_superblock_test 00:13:55.774 ************************************ 00:13:56.034 15:18:23 bdev_raid -- bdev/bdev_raid.sh@989 -- # '[' true = true ']' 00:13:56.035 15:18:23 bdev_raid -- bdev/bdev_raid.sh@990 -- # run_test raid5f_rebuild_test raid_rebuild_test raid5f 3 false false true 00:13:56.035 15:18:23 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:13:56.035 15:18:23 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:56.035 15:18:23 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:56.035 ************************************ 00:13:56.035 START TEST raid5f_rebuild_test 00:13:56.035 ************************************ 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid5f 3 false false true 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=3 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=92270 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 92270 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@835 -- # '[' -z 92270 ']' 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:56.035 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:13:56.035 15:18:23 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:56.035 [2024-11-27 15:18:24.047457] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:13:56.035 [2024-11-27 15:18:24.047712] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid92270 ] 00:13:56.035 I/O size of 3145728 is greater than zero copy threshold (65536). 00:13:56.035 Zero copy mechanism will not be used. 00:13:56.295 [2024-11-27 15:18:24.223643] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:56.295 [2024-11-27 15:18:24.266504] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:56.295 [2024-11-27 15:18:24.343886] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:56.295 [2024-11-27 15:18:24.344036] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:56.865 15:18:24 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:13:56.865 15:18:24 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # return 0 00:13:56.865 15:18:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:56.866 15:18:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:13:56.866 15:18:24 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:56.866 15:18:24 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:56.866 BaseBdev1_malloc 00:13:56.866 15:18:24 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:56.866 15:18:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:56.866 15:18:24 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:56.866 15:18:24 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:56.866 [2024-11-27 15:18:24.925262] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:56.866 [2024-11-27 15:18:24.925421] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:56.866 [2024-11-27 15:18:24.925481] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:13:56.866 [2024-11-27 15:18:24.925526] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:56.866 [2024-11-27 15:18:24.927827] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:56.866 BaseBdev1 00:13:56.866 [2024-11-27 15:18:24.927926] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:56.866 15:18:24 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:56.866 15:18:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:56.866 15:18:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:13:56.866 15:18:24 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:56.866 15:18:24 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:56.866 BaseBdev2_malloc 00:13:56.866 15:18:24 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:56.866 15:18:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:13:56.866 15:18:24 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:56.866 15:18:24 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:56.866 [2024-11-27 15:18:24.959689] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:13:56.866 [2024-11-27 15:18:24.959797] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:56.866 [2024-11-27 15:18:24.959840] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:13:56.866 [2024-11-27 15:18:24.959874] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:56.866 [2024-11-27 15:18:24.962224] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:56.866 [2024-11-27 15:18:24.962304] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:13:56.866 BaseBdev2 00:13:56.866 15:18:24 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:56.866 15:18:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:56.866 15:18:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:13:56.866 15:18:24 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:56.866 15:18:24 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:57.125 BaseBdev3_malloc 00:13:57.125 15:18:24 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:57.125 15:18:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:13:57.125 15:18:24 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:57.125 15:18:24 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:57.125 [2024-11-27 15:18:24.994177] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:13:57.125 [2024-11-27 15:18:24.994287] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:57.125 [2024-11-27 15:18:24.994335] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:13:57.125 [2024-11-27 15:18:24.994368] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:57.125 [2024-11-27 15:18:24.996667] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:57.125 [2024-11-27 15:18:24.996747] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:13:57.125 BaseBdev3 00:13:57.125 15:18:24 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:57.125 15:18:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:13:57.125 15:18:24 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:57.125 15:18:24 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:57.125 spare_malloc 00:13:57.125 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:57.125 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:13:57.125 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:57.125 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:57.125 spare_delay 00:13:57.125 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:57.125 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:57.125 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:57.125 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:57.125 [2024-11-27 15:18:25.048256] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:57.125 [2024-11-27 15:18:25.048360] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:57.125 [2024-11-27 15:18:25.048405] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:13:57.125 [2024-11-27 15:18:25.048438] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:57.125 [2024-11-27 15:18:25.050728] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:57.125 [2024-11-27 15:18:25.050812] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:57.125 spare 00:13:57.125 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:57.125 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 00:13:57.125 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:57.125 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:57.126 [2024-11-27 15:18:25.060292] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:57.126 [2024-11-27 15:18:25.062324] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:57.126 [2024-11-27 15:18:25.062437] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:57.126 [2024-11-27 15:18:25.062543] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:13:57.126 [2024-11-27 15:18:25.062589] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:13:57.126 [2024-11-27 15:18:25.062858] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:13:57.126 [2024-11-27 15:18:25.063364] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:13:57.126 [2024-11-27 15:18:25.063429] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:13:57.126 [2024-11-27 15:18:25.063590] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:57.126 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:57.126 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:57.126 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:57.126 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:57.126 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:57.126 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:57.126 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:57.126 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:57.126 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:57.126 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:57.126 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:57.126 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:57.126 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:57.126 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:57.126 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:57.126 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:57.126 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:57.126 "name": "raid_bdev1", 00:13:57.126 "uuid": "064cbbee-c103-40f7-91eb-a8337db040d8", 00:13:57.126 "strip_size_kb": 64, 00:13:57.126 "state": "online", 00:13:57.126 "raid_level": "raid5f", 00:13:57.126 "superblock": false, 00:13:57.126 "num_base_bdevs": 3, 00:13:57.126 "num_base_bdevs_discovered": 3, 00:13:57.126 "num_base_bdevs_operational": 3, 00:13:57.126 "base_bdevs_list": [ 00:13:57.126 { 00:13:57.126 "name": "BaseBdev1", 00:13:57.126 "uuid": "b6d40527-340d-59c3-8cf0-d1d94b400e1d", 00:13:57.126 "is_configured": true, 00:13:57.126 "data_offset": 0, 00:13:57.126 "data_size": 65536 00:13:57.126 }, 00:13:57.126 { 00:13:57.126 "name": "BaseBdev2", 00:13:57.126 "uuid": "57e3e4da-d7e9-5f72-a4cb-67e4e546e4d4", 00:13:57.126 "is_configured": true, 00:13:57.126 "data_offset": 0, 00:13:57.126 "data_size": 65536 00:13:57.126 }, 00:13:57.126 { 00:13:57.126 "name": "BaseBdev3", 00:13:57.126 "uuid": "567f54ce-90d2-5fc2-88ce-3dcbaabc8155", 00:13:57.126 "is_configured": true, 00:13:57.126 "data_offset": 0, 00:13:57.126 "data_size": 65536 00:13:57.126 } 00:13:57.126 ] 00:13:57.126 }' 00:13:57.126 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:57.126 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:57.694 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:57.694 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:57.694 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:57.694 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:13:57.694 [2024-11-27 15:18:25.533359] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:57.694 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:57.694 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=131072 00:13:57.694 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:57.694 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:13:57.694 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:57.694 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:57.694 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:57.694 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:13:57.694 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:13:57.694 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:13:57.694 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:13:57.694 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:13:57.694 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:57.694 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:13:57.694 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:57.694 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:13:57.694 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:57.694 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:13:57.694 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:57.694 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:57.694 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:13:57.953 [2024-11-27 15:18:25.816758] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:13:57.953 /dev/nbd0 00:13:57.953 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:57.953 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:57.953 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:13:57.953 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:13:57.953 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:57.954 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:57.954 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:13:57.954 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:13:57.954 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:57.954 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:57.954 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:57.954 1+0 records in 00:13:57.954 1+0 records out 00:13:57.954 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000552847 s, 7.4 MB/s 00:13:57.954 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:57.954 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:13:57.954 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:57.954 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:57.954 15:18:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:13:57.954 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:57.954 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:57.954 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:13:57.954 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@630 -- # write_unit_size=256 00:13:57.954 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@631 -- # echo 128 00:13:57.954 15:18:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=131072 count=512 oflag=direct 00:13:58.214 512+0 records in 00:13:58.214 512+0 records out 00:13:58.214 67108864 bytes (67 MB, 64 MiB) copied, 0.291534 s, 230 MB/s 00:13:58.214 15:18:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:13:58.214 15:18:26 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:58.214 15:18:26 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:13:58.214 15:18:26 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:58.214 15:18:26 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:13:58.214 15:18:26 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:58.214 15:18:26 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:58.474 15:18:26 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:58.474 [2024-11-27 15:18:26.414260] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:58.474 15:18:26 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:58.474 15:18:26 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:58.474 15:18:26 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:58.474 15:18:26 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:58.474 15:18:26 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:58.474 15:18:26 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:13:58.474 15:18:26 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:13:58.474 15:18:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:13:58.474 15:18:26 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:58.474 15:18:26 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:58.474 [2024-11-27 15:18:26.432346] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:58.474 15:18:26 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:58.474 15:18:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:58.474 15:18:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:58.474 15:18:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:58.474 15:18:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:58.474 15:18:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:58.474 15:18:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:58.474 15:18:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:58.474 15:18:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:58.474 15:18:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:58.474 15:18:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:58.474 15:18:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:58.474 15:18:26 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:58.474 15:18:26 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:58.474 15:18:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:58.474 15:18:26 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:58.474 15:18:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:58.474 "name": "raid_bdev1", 00:13:58.474 "uuid": "064cbbee-c103-40f7-91eb-a8337db040d8", 00:13:58.474 "strip_size_kb": 64, 00:13:58.474 "state": "online", 00:13:58.474 "raid_level": "raid5f", 00:13:58.474 "superblock": false, 00:13:58.474 "num_base_bdevs": 3, 00:13:58.474 "num_base_bdevs_discovered": 2, 00:13:58.474 "num_base_bdevs_operational": 2, 00:13:58.474 "base_bdevs_list": [ 00:13:58.474 { 00:13:58.474 "name": null, 00:13:58.474 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:58.474 "is_configured": false, 00:13:58.474 "data_offset": 0, 00:13:58.474 "data_size": 65536 00:13:58.474 }, 00:13:58.474 { 00:13:58.474 "name": "BaseBdev2", 00:13:58.474 "uuid": "57e3e4da-d7e9-5f72-a4cb-67e4e546e4d4", 00:13:58.474 "is_configured": true, 00:13:58.474 "data_offset": 0, 00:13:58.474 "data_size": 65536 00:13:58.474 }, 00:13:58.474 { 00:13:58.474 "name": "BaseBdev3", 00:13:58.474 "uuid": "567f54ce-90d2-5fc2-88ce-3dcbaabc8155", 00:13:58.474 "is_configured": true, 00:13:58.474 "data_offset": 0, 00:13:58.474 "data_size": 65536 00:13:58.474 } 00:13:58.474 ] 00:13:58.474 }' 00:13:58.474 15:18:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:58.474 15:18:26 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:59.044 15:18:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:59.044 15:18:26 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:59.044 15:18:26 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:59.044 [2024-11-27 15:18:26.863639] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:59.044 [2024-11-27 15:18:26.868644] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b4e0 00:13:59.044 15:18:26 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:59.044 15:18:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:13:59.044 [2024-11-27 15:18:26.871051] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:59.982 15:18:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:59.982 15:18:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:59.982 15:18:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:59.982 15:18:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:59.982 15:18:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:59.982 15:18:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:59.982 15:18:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:59.982 15:18:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:59.982 15:18:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:59.982 15:18:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:59.982 15:18:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:59.982 "name": "raid_bdev1", 00:13:59.982 "uuid": "064cbbee-c103-40f7-91eb-a8337db040d8", 00:13:59.982 "strip_size_kb": 64, 00:13:59.982 "state": "online", 00:13:59.982 "raid_level": "raid5f", 00:13:59.982 "superblock": false, 00:13:59.982 "num_base_bdevs": 3, 00:13:59.982 "num_base_bdevs_discovered": 3, 00:13:59.982 "num_base_bdevs_operational": 3, 00:13:59.982 "process": { 00:13:59.982 "type": "rebuild", 00:13:59.982 "target": "spare", 00:13:59.982 "progress": { 00:13:59.982 "blocks": 20480, 00:13:59.982 "percent": 15 00:13:59.982 } 00:13:59.982 }, 00:13:59.983 "base_bdevs_list": [ 00:13:59.983 { 00:13:59.983 "name": "spare", 00:13:59.983 "uuid": "d4417e6e-a6f4-5ec6-b97d-fef5da88a7e3", 00:13:59.983 "is_configured": true, 00:13:59.983 "data_offset": 0, 00:13:59.983 "data_size": 65536 00:13:59.983 }, 00:13:59.983 { 00:13:59.983 "name": "BaseBdev2", 00:13:59.983 "uuid": "57e3e4da-d7e9-5f72-a4cb-67e4e546e4d4", 00:13:59.983 "is_configured": true, 00:13:59.983 "data_offset": 0, 00:13:59.983 "data_size": 65536 00:13:59.983 }, 00:13:59.983 { 00:13:59.983 "name": "BaseBdev3", 00:13:59.983 "uuid": "567f54ce-90d2-5fc2-88ce-3dcbaabc8155", 00:13:59.983 "is_configured": true, 00:13:59.983 "data_offset": 0, 00:13:59.983 "data_size": 65536 00:13:59.983 } 00:13:59.983 ] 00:13:59.983 }' 00:13:59.983 15:18:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:59.983 15:18:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:59.983 15:18:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:59.983 15:18:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:59.983 15:18:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:59.983 15:18:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:59.983 15:18:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:59.983 [2024-11-27 15:18:28.006969] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:59.983 [2024-11-27 15:18:28.079596] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:59.983 [2024-11-27 15:18:28.079674] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:59.983 [2024-11-27 15:18:28.079691] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:59.983 [2024-11-27 15:18:28.079701] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:00.242 15:18:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:00.242 15:18:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:00.242 15:18:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:00.242 15:18:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:00.242 15:18:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:00.242 15:18:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:00.242 15:18:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:00.242 15:18:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:00.242 15:18:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:00.242 15:18:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:00.242 15:18:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:00.242 15:18:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:00.242 15:18:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:00.242 15:18:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:00.242 15:18:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:00.242 15:18:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:00.242 15:18:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:00.242 "name": "raid_bdev1", 00:14:00.243 "uuid": "064cbbee-c103-40f7-91eb-a8337db040d8", 00:14:00.243 "strip_size_kb": 64, 00:14:00.243 "state": "online", 00:14:00.243 "raid_level": "raid5f", 00:14:00.243 "superblock": false, 00:14:00.243 "num_base_bdevs": 3, 00:14:00.243 "num_base_bdevs_discovered": 2, 00:14:00.243 "num_base_bdevs_operational": 2, 00:14:00.243 "base_bdevs_list": [ 00:14:00.243 { 00:14:00.243 "name": null, 00:14:00.243 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:00.243 "is_configured": false, 00:14:00.243 "data_offset": 0, 00:14:00.243 "data_size": 65536 00:14:00.243 }, 00:14:00.243 { 00:14:00.243 "name": "BaseBdev2", 00:14:00.243 "uuid": "57e3e4da-d7e9-5f72-a4cb-67e4e546e4d4", 00:14:00.243 "is_configured": true, 00:14:00.243 "data_offset": 0, 00:14:00.243 "data_size": 65536 00:14:00.243 }, 00:14:00.243 { 00:14:00.243 "name": "BaseBdev3", 00:14:00.243 "uuid": "567f54ce-90d2-5fc2-88ce-3dcbaabc8155", 00:14:00.243 "is_configured": true, 00:14:00.243 "data_offset": 0, 00:14:00.243 "data_size": 65536 00:14:00.243 } 00:14:00.243 ] 00:14:00.243 }' 00:14:00.243 15:18:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:00.243 15:18:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:00.503 15:18:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:00.503 15:18:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:00.503 15:18:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:00.503 15:18:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:00.503 15:18:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:00.503 15:18:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:00.503 15:18:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:00.503 15:18:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:00.503 15:18:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:00.503 15:18:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:00.763 15:18:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:00.763 "name": "raid_bdev1", 00:14:00.763 "uuid": "064cbbee-c103-40f7-91eb-a8337db040d8", 00:14:00.763 "strip_size_kb": 64, 00:14:00.763 "state": "online", 00:14:00.763 "raid_level": "raid5f", 00:14:00.763 "superblock": false, 00:14:00.763 "num_base_bdevs": 3, 00:14:00.763 "num_base_bdevs_discovered": 2, 00:14:00.763 "num_base_bdevs_operational": 2, 00:14:00.763 "base_bdevs_list": [ 00:14:00.763 { 00:14:00.763 "name": null, 00:14:00.763 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:00.763 "is_configured": false, 00:14:00.763 "data_offset": 0, 00:14:00.763 "data_size": 65536 00:14:00.763 }, 00:14:00.763 { 00:14:00.763 "name": "BaseBdev2", 00:14:00.763 "uuid": "57e3e4da-d7e9-5f72-a4cb-67e4e546e4d4", 00:14:00.763 "is_configured": true, 00:14:00.763 "data_offset": 0, 00:14:00.763 "data_size": 65536 00:14:00.763 }, 00:14:00.763 { 00:14:00.763 "name": "BaseBdev3", 00:14:00.763 "uuid": "567f54ce-90d2-5fc2-88ce-3dcbaabc8155", 00:14:00.763 "is_configured": true, 00:14:00.763 "data_offset": 0, 00:14:00.763 "data_size": 65536 00:14:00.763 } 00:14:00.763 ] 00:14:00.763 }' 00:14:00.763 15:18:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:00.763 15:18:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:00.763 15:18:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:00.763 15:18:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:00.763 15:18:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:00.763 15:18:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:00.763 15:18:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:00.763 [2024-11-27 15:18:28.692901] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:00.763 [2024-11-27 15:18:28.697637] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b5b0 00:14:00.763 15:18:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:00.763 15:18:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:14:00.763 [2024-11-27 15:18:28.700089] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:01.703 15:18:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:01.703 15:18:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:01.703 15:18:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:01.703 15:18:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:01.703 15:18:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:01.703 15:18:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:01.703 15:18:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:01.703 15:18:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.703 15:18:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:01.703 15:18:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.703 15:18:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:01.703 "name": "raid_bdev1", 00:14:01.703 "uuid": "064cbbee-c103-40f7-91eb-a8337db040d8", 00:14:01.703 "strip_size_kb": 64, 00:14:01.703 "state": "online", 00:14:01.703 "raid_level": "raid5f", 00:14:01.703 "superblock": false, 00:14:01.703 "num_base_bdevs": 3, 00:14:01.703 "num_base_bdevs_discovered": 3, 00:14:01.703 "num_base_bdevs_operational": 3, 00:14:01.703 "process": { 00:14:01.703 "type": "rebuild", 00:14:01.703 "target": "spare", 00:14:01.703 "progress": { 00:14:01.703 "blocks": 20480, 00:14:01.703 "percent": 15 00:14:01.703 } 00:14:01.703 }, 00:14:01.703 "base_bdevs_list": [ 00:14:01.703 { 00:14:01.703 "name": "spare", 00:14:01.703 "uuid": "d4417e6e-a6f4-5ec6-b97d-fef5da88a7e3", 00:14:01.703 "is_configured": true, 00:14:01.703 "data_offset": 0, 00:14:01.703 "data_size": 65536 00:14:01.703 }, 00:14:01.703 { 00:14:01.703 "name": "BaseBdev2", 00:14:01.703 "uuid": "57e3e4da-d7e9-5f72-a4cb-67e4e546e4d4", 00:14:01.703 "is_configured": true, 00:14:01.703 "data_offset": 0, 00:14:01.703 "data_size": 65536 00:14:01.703 }, 00:14:01.703 { 00:14:01.703 "name": "BaseBdev3", 00:14:01.703 "uuid": "567f54ce-90d2-5fc2-88ce-3dcbaabc8155", 00:14:01.703 "is_configured": true, 00:14:01.703 "data_offset": 0, 00:14:01.703 "data_size": 65536 00:14:01.703 } 00:14:01.703 ] 00:14:01.703 }' 00:14:01.703 15:18:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:01.963 15:18:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:01.963 15:18:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:01.963 15:18:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:01.963 15:18:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:14:01.963 15:18:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=3 00:14:01.963 15:18:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:14:01.963 15:18:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=450 00:14:01.963 15:18:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:01.963 15:18:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:01.963 15:18:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:01.963 15:18:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:01.963 15:18:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:01.963 15:18:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:01.963 15:18:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:01.963 15:18:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:01.963 15:18:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.963 15:18:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:01.963 15:18:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.963 15:18:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:01.963 "name": "raid_bdev1", 00:14:01.963 "uuid": "064cbbee-c103-40f7-91eb-a8337db040d8", 00:14:01.963 "strip_size_kb": 64, 00:14:01.963 "state": "online", 00:14:01.963 "raid_level": "raid5f", 00:14:01.963 "superblock": false, 00:14:01.963 "num_base_bdevs": 3, 00:14:01.963 "num_base_bdevs_discovered": 3, 00:14:01.963 "num_base_bdevs_operational": 3, 00:14:01.963 "process": { 00:14:01.963 "type": "rebuild", 00:14:01.963 "target": "spare", 00:14:01.963 "progress": { 00:14:01.963 "blocks": 22528, 00:14:01.963 "percent": 17 00:14:01.963 } 00:14:01.963 }, 00:14:01.963 "base_bdevs_list": [ 00:14:01.963 { 00:14:01.963 "name": "spare", 00:14:01.963 "uuid": "d4417e6e-a6f4-5ec6-b97d-fef5da88a7e3", 00:14:01.963 "is_configured": true, 00:14:01.963 "data_offset": 0, 00:14:01.963 "data_size": 65536 00:14:01.963 }, 00:14:01.963 { 00:14:01.963 "name": "BaseBdev2", 00:14:01.963 "uuid": "57e3e4da-d7e9-5f72-a4cb-67e4e546e4d4", 00:14:01.963 "is_configured": true, 00:14:01.963 "data_offset": 0, 00:14:01.963 "data_size": 65536 00:14:01.963 }, 00:14:01.963 { 00:14:01.963 "name": "BaseBdev3", 00:14:01.963 "uuid": "567f54ce-90d2-5fc2-88ce-3dcbaabc8155", 00:14:01.963 "is_configured": true, 00:14:01.963 "data_offset": 0, 00:14:01.963 "data_size": 65536 00:14:01.963 } 00:14:01.963 ] 00:14:01.963 }' 00:14:01.963 15:18:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:01.963 15:18:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:01.963 15:18:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:01.963 15:18:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:01.963 15:18:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:02.902 15:18:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:02.902 15:18:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:02.902 15:18:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:02.902 15:18:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:02.902 15:18:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:02.902 15:18:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:02.902 15:18:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:02.902 15:18:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:02.902 15:18:31 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:02.902 15:18:31 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:03.162 15:18:31 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:03.162 15:18:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:03.162 "name": "raid_bdev1", 00:14:03.162 "uuid": "064cbbee-c103-40f7-91eb-a8337db040d8", 00:14:03.162 "strip_size_kb": 64, 00:14:03.162 "state": "online", 00:14:03.162 "raid_level": "raid5f", 00:14:03.162 "superblock": false, 00:14:03.162 "num_base_bdevs": 3, 00:14:03.162 "num_base_bdevs_discovered": 3, 00:14:03.162 "num_base_bdevs_operational": 3, 00:14:03.162 "process": { 00:14:03.162 "type": "rebuild", 00:14:03.162 "target": "spare", 00:14:03.162 "progress": { 00:14:03.162 "blocks": 45056, 00:14:03.162 "percent": 34 00:14:03.162 } 00:14:03.162 }, 00:14:03.162 "base_bdevs_list": [ 00:14:03.162 { 00:14:03.162 "name": "spare", 00:14:03.162 "uuid": "d4417e6e-a6f4-5ec6-b97d-fef5da88a7e3", 00:14:03.162 "is_configured": true, 00:14:03.162 "data_offset": 0, 00:14:03.162 "data_size": 65536 00:14:03.162 }, 00:14:03.162 { 00:14:03.162 "name": "BaseBdev2", 00:14:03.162 "uuid": "57e3e4da-d7e9-5f72-a4cb-67e4e546e4d4", 00:14:03.162 "is_configured": true, 00:14:03.162 "data_offset": 0, 00:14:03.162 "data_size": 65536 00:14:03.162 }, 00:14:03.162 { 00:14:03.162 "name": "BaseBdev3", 00:14:03.162 "uuid": "567f54ce-90d2-5fc2-88ce-3dcbaabc8155", 00:14:03.162 "is_configured": true, 00:14:03.162 "data_offset": 0, 00:14:03.162 "data_size": 65536 00:14:03.162 } 00:14:03.162 ] 00:14:03.162 }' 00:14:03.162 15:18:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:03.162 15:18:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:03.162 15:18:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:03.162 15:18:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:03.162 15:18:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:04.101 15:18:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:04.101 15:18:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:04.101 15:18:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:04.101 15:18:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:04.101 15:18:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:04.101 15:18:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:04.101 15:18:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:04.101 15:18:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:04.101 15:18:32 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.101 15:18:32 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.101 15:18:32 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.101 15:18:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:04.101 "name": "raid_bdev1", 00:14:04.101 "uuid": "064cbbee-c103-40f7-91eb-a8337db040d8", 00:14:04.101 "strip_size_kb": 64, 00:14:04.101 "state": "online", 00:14:04.101 "raid_level": "raid5f", 00:14:04.101 "superblock": false, 00:14:04.101 "num_base_bdevs": 3, 00:14:04.101 "num_base_bdevs_discovered": 3, 00:14:04.101 "num_base_bdevs_operational": 3, 00:14:04.101 "process": { 00:14:04.101 "type": "rebuild", 00:14:04.101 "target": "spare", 00:14:04.102 "progress": { 00:14:04.102 "blocks": 69632, 00:14:04.102 "percent": 53 00:14:04.102 } 00:14:04.102 }, 00:14:04.102 "base_bdevs_list": [ 00:14:04.102 { 00:14:04.102 "name": "spare", 00:14:04.102 "uuid": "d4417e6e-a6f4-5ec6-b97d-fef5da88a7e3", 00:14:04.102 "is_configured": true, 00:14:04.102 "data_offset": 0, 00:14:04.102 "data_size": 65536 00:14:04.102 }, 00:14:04.102 { 00:14:04.102 "name": "BaseBdev2", 00:14:04.102 "uuid": "57e3e4da-d7e9-5f72-a4cb-67e4e546e4d4", 00:14:04.102 "is_configured": true, 00:14:04.102 "data_offset": 0, 00:14:04.102 "data_size": 65536 00:14:04.102 }, 00:14:04.102 { 00:14:04.102 "name": "BaseBdev3", 00:14:04.102 "uuid": "567f54ce-90d2-5fc2-88ce-3dcbaabc8155", 00:14:04.102 "is_configured": true, 00:14:04.102 "data_offset": 0, 00:14:04.102 "data_size": 65536 00:14:04.102 } 00:14:04.102 ] 00:14:04.102 }' 00:14:04.102 15:18:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:04.361 15:18:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:04.361 15:18:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:04.361 15:18:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:04.361 15:18:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:05.357 15:18:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:05.357 15:18:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:05.357 15:18:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:05.357 15:18:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:05.357 15:18:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:05.357 15:18:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:05.357 15:18:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:05.357 15:18:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:05.357 15:18:33 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:05.357 15:18:33 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:05.357 15:18:33 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:05.357 15:18:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:05.357 "name": "raid_bdev1", 00:14:05.357 "uuid": "064cbbee-c103-40f7-91eb-a8337db040d8", 00:14:05.357 "strip_size_kb": 64, 00:14:05.357 "state": "online", 00:14:05.357 "raid_level": "raid5f", 00:14:05.357 "superblock": false, 00:14:05.357 "num_base_bdevs": 3, 00:14:05.357 "num_base_bdevs_discovered": 3, 00:14:05.357 "num_base_bdevs_operational": 3, 00:14:05.357 "process": { 00:14:05.357 "type": "rebuild", 00:14:05.357 "target": "spare", 00:14:05.357 "progress": { 00:14:05.357 "blocks": 92160, 00:14:05.357 "percent": 70 00:14:05.357 } 00:14:05.357 }, 00:14:05.357 "base_bdevs_list": [ 00:14:05.357 { 00:14:05.357 "name": "spare", 00:14:05.357 "uuid": "d4417e6e-a6f4-5ec6-b97d-fef5da88a7e3", 00:14:05.357 "is_configured": true, 00:14:05.357 "data_offset": 0, 00:14:05.357 "data_size": 65536 00:14:05.357 }, 00:14:05.357 { 00:14:05.357 "name": "BaseBdev2", 00:14:05.357 "uuid": "57e3e4da-d7e9-5f72-a4cb-67e4e546e4d4", 00:14:05.357 "is_configured": true, 00:14:05.357 "data_offset": 0, 00:14:05.357 "data_size": 65536 00:14:05.357 }, 00:14:05.357 { 00:14:05.357 "name": "BaseBdev3", 00:14:05.357 "uuid": "567f54ce-90d2-5fc2-88ce-3dcbaabc8155", 00:14:05.357 "is_configured": true, 00:14:05.357 "data_offset": 0, 00:14:05.357 "data_size": 65536 00:14:05.357 } 00:14:05.357 ] 00:14:05.357 }' 00:14:05.357 15:18:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:05.357 15:18:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:05.357 15:18:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:05.357 15:18:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:05.357 15:18:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:06.739 15:18:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:06.739 15:18:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:06.739 15:18:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:06.739 15:18:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:06.739 15:18:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:06.739 15:18:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:06.739 15:18:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:06.739 15:18:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:06.739 15:18:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:06.739 15:18:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:06.739 15:18:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:06.739 15:18:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:06.739 "name": "raid_bdev1", 00:14:06.739 "uuid": "064cbbee-c103-40f7-91eb-a8337db040d8", 00:14:06.739 "strip_size_kb": 64, 00:14:06.739 "state": "online", 00:14:06.739 "raid_level": "raid5f", 00:14:06.739 "superblock": false, 00:14:06.739 "num_base_bdevs": 3, 00:14:06.739 "num_base_bdevs_discovered": 3, 00:14:06.739 "num_base_bdevs_operational": 3, 00:14:06.739 "process": { 00:14:06.739 "type": "rebuild", 00:14:06.739 "target": "spare", 00:14:06.739 "progress": { 00:14:06.739 "blocks": 114688, 00:14:06.739 "percent": 87 00:14:06.739 } 00:14:06.739 }, 00:14:06.739 "base_bdevs_list": [ 00:14:06.739 { 00:14:06.739 "name": "spare", 00:14:06.739 "uuid": "d4417e6e-a6f4-5ec6-b97d-fef5da88a7e3", 00:14:06.739 "is_configured": true, 00:14:06.739 "data_offset": 0, 00:14:06.739 "data_size": 65536 00:14:06.739 }, 00:14:06.739 { 00:14:06.739 "name": "BaseBdev2", 00:14:06.739 "uuid": "57e3e4da-d7e9-5f72-a4cb-67e4e546e4d4", 00:14:06.739 "is_configured": true, 00:14:06.739 "data_offset": 0, 00:14:06.739 "data_size": 65536 00:14:06.739 }, 00:14:06.739 { 00:14:06.739 "name": "BaseBdev3", 00:14:06.739 "uuid": "567f54ce-90d2-5fc2-88ce-3dcbaabc8155", 00:14:06.739 "is_configured": true, 00:14:06.739 "data_offset": 0, 00:14:06.739 "data_size": 65536 00:14:06.739 } 00:14:06.739 ] 00:14:06.739 }' 00:14:06.739 15:18:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:06.739 15:18:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:06.739 15:18:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:06.739 15:18:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:06.739 15:18:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:07.310 [2024-11-27 15:18:35.144569] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:14:07.310 [2024-11-27 15:18:35.144760] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:14:07.310 [2024-11-27 15:18:35.144848] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:07.603 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:07.603 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:07.603 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:07.603 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:07.603 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:07.603 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:07.603 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:07.603 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:07.603 15:18:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:07.603 15:18:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:07.603 15:18:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:07.603 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:07.603 "name": "raid_bdev1", 00:14:07.603 "uuid": "064cbbee-c103-40f7-91eb-a8337db040d8", 00:14:07.603 "strip_size_kb": 64, 00:14:07.603 "state": "online", 00:14:07.603 "raid_level": "raid5f", 00:14:07.603 "superblock": false, 00:14:07.603 "num_base_bdevs": 3, 00:14:07.603 "num_base_bdevs_discovered": 3, 00:14:07.603 "num_base_bdevs_operational": 3, 00:14:07.603 "base_bdevs_list": [ 00:14:07.603 { 00:14:07.603 "name": "spare", 00:14:07.603 "uuid": "d4417e6e-a6f4-5ec6-b97d-fef5da88a7e3", 00:14:07.603 "is_configured": true, 00:14:07.603 "data_offset": 0, 00:14:07.603 "data_size": 65536 00:14:07.603 }, 00:14:07.603 { 00:14:07.603 "name": "BaseBdev2", 00:14:07.603 "uuid": "57e3e4da-d7e9-5f72-a4cb-67e4e546e4d4", 00:14:07.603 "is_configured": true, 00:14:07.603 "data_offset": 0, 00:14:07.603 "data_size": 65536 00:14:07.603 }, 00:14:07.603 { 00:14:07.603 "name": "BaseBdev3", 00:14:07.603 "uuid": "567f54ce-90d2-5fc2-88ce-3dcbaabc8155", 00:14:07.603 "is_configured": true, 00:14:07.603 "data_offset": 0, 00:14:07.603 "data_size": 65536 00:14:07.603 } 00:14:07.603 ] 00:14:07.603 }' 00:14:07.603 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:07.603 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:14:07.603 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:07.863 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:14:07.863 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:14:07.863 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:07.863 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:07.863 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:07.863 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:07.863 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:07.863 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:07.864 15:18:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:07.864 15:18:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:07.864 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:07.864 15:18:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:07.864 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:07.864 "name": "raid_bdev1", 00:14:07.864 "uuid": "064cbbee-c103-40f7-91eb-a8337db040d8", 00:14:07.864 "strip_size_kb": 64, 00:14:07.864 "state": "online", 00:14:07.864 "raid_level": "raid5f", 00:14:07.864 "superblock": false, 00:14:07.864 "num_base_bdevs": 3, 00:14:07.864 "num_base_bdevs_discovered": 3, 00:14:07.864 "num_base_bdevs_operational": 3, 00:14:07.864 "base_bdevs_list": [ 00:14:07.864 { 00:14:07.864 "name": "spare", 00:14:07.864 "uuid": "d4417e6e-a6f4-5ec6-b97d-fef5da88a7e3", 00:14:07.864 "is_configured": true, 00:14:07.864 "data_offset": 0, 00:14:07.864 "data_size": 65536 00:14:07.864 }, 00:14:07.864 { 00:14:07.864 "name": "BaseBdev2", 00:14:07.864 "uuid": "57e3e4da-d7e9-5f72-a4cb-67e4e546e4d4", 00:14:07.864 "is_configured": true, 00:14:07.864 "data_offset": 0, 00:14:07.864 "data_size": 65536 00:14:07.864 }, 00:14:07.864 { 00:14:07.864 "name": "BaseBdev3", 00:14:07.864 "uuid": "567f54ce-90d2-5fc2-88ce-3dcbaabc8155", 00:14:07.864 "is_configured": true, 00:14:07.864 "data_offset": 0, 00:14:07.864 "data_size": 65536 00:14:07.864 } 00:14:07.864 ] 00:14:07.864 }' 00:14:07.864 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:07.864 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:07.864 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:07.864 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:07.864 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:07.864 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:07.864 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:07.864 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:07.864 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:07.864 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:07.864 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:07.864 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:07.864 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:07.864 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:07.864 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:07.864 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:07.864 15:18:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:07.864 15:18:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:07.864 15:18:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:07.864 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:07.864 "name": "raid_bdev1", 00:14:07.864 "uuid": "064cbbee-c103-40f7-91eb-a8337db040d8", 00:14:07.864 "strip_size_kb": 64, 00:14:07.864 "state": "online", 00:14:07.864 "raid_level": "raid5f", 00:14:07.864 "superblock": false, 00:14:07.864 "num_base_bdevs": 3, 00:14:07.864 "num_base_bdevs_discovered": 3, 00:14:07.864 "num_base_bdevs_operational": 3, 00:14:07.864 "base_bdevs_list": [ 00:14:07.864 { 00:14:07.864 "name": "spare", 00:14:07.864 "uuid": "d4417e6e-a6f4-5ec6-b97d-fef5da88a7e3", 00:14:07.864 "is_configured": true, 00:14:07.864 "data_offset": 0, 00:14:07.864 "data_size": 65536 00:14:07.864 }, 00:14:07.864 { 00:14:07.864 "name": "BaseBdev2", 00:14:07.864 "uuid": "57e3e4da-d7e9-5f72-a4cb-67e4e546e4d4", 00:14:07.864 "is_configured": true, 00:14:07.864 "data_offset": 0, 00:14:07.864 "data_size": 65536 00:14:07.864 }, 00:14:07.864 { 00:14:07.864 "name": "BaseBdev3", 00:14:07.864 "uuid": "567f54ce-90d2-5fc2-88ce-3dcbaabc8155", 00:14:07.864 "is_configured": true, 00:14:07.864 "data_offset": 0, 00:14:07.864 "data_size": 65536 00:14:07.864 } 00:14:07.864 ] 00:14:07.864 }' 00:14:07.864 15:18:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:07.864 15:18:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.434 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:08.434 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:08.434 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.434 [2024-11-27 15:18:36.296816] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:08.434 [2024-11-27 15:18:36.296914] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:08.434 [2024-11-27 15:18:36.297028] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:08.434 [2024-11-27 15:18:36.297161] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:08.434 [2024-11-27 15:18:36.297222] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:14:08.434 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:08.434 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:08.434 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:08.434 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.434 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:14:08.434 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:08.434 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:14:08.434 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:14:08.434 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:14:08.434 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:14:08.434 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:08.434 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:14:08.434 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:08.434 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:14:08.434 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:08.434 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:14:08.434 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:08.434 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:08.434 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:14:08.697 /dev/nbd0 00:14:08.697 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:08.697 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:08.697 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:14:08.697 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:14:08.697 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:08.697 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:08.697 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:14:08.697 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:14:08.697 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:08.697 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:08.697 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:08.697 1+0 records in 00:14:08.697 1+0 records out 00:14:08.697 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000483203 s, 8.5 MB/s 00:14:08.697 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:08.697 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:14:08.697 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:08.697 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:08.697 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:14:08.697 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:08.697 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:08.697 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:14:08.955 /dev/nbd1 00:14:08.955 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:14:08.955 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:14:08.955 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:14:08.956 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:14:08.956 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:08.956 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:08.956 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:14:08.956 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:14:08.956 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:08.956 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:08.956 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:08.956 1+0 records in 00:14:08.956 1+0 records out 00:14:08.956 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000405784 s, 10.1 MB/s 00:14:08.956 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:08.956 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:14:08.956 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:08.956 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:08.956 15:18:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:14:08.956 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:08.956 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:08.956 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:14:08.956 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:14:08.956 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:08.956 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:14:08.956 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:08.956 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:14:08.956 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:08.956 15:18:36 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:09.215 15:18:37 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:09.215 15:18:37 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:09.215 15:18:37 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:09.215 15:18:37 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:09.215 15:18:37 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:09.215 15:18:37 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:09.215 15:18:37 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:14:09.215 15:18:37 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:14:09.215 15:18:37 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:09.215 15:18:37 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:14:09.475 15:18:37 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:14:09.475 15:18:37 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:14:09.475 15:18:37 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:14:09.475 15:18:37 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:09.475 15:18:37 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:09.475 15:18:37 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:14:09.475 15:18:37 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:14:09.475 15:18:37 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:14:09.475 15:18:37 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:14:09.475 15:18:37 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 92270 00:14:09.475 15:18:37 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@954 -- # '[' -z 92270 ']' 00:14:09.475 15:18:37 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@958 -- # kill -0 92270 00:14:09.475 15:18:37 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@959 -- # uname 00:14:09.475 15:18:37 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:09.475 15:18:37 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 92270 00:14:09.475 killing process with pid 92270 00:14:09.475 Received shutdown signal, test time was about 60.000000 seconds 00:14:09.475 00:14:09.475 Latency(us) 00:14:09.475 [2024-11-27T15:18:37.582Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:14:09.475 [2024-11-27T15:18:37.582Z] =================================================================================================================== 00:14:09.475 [2024-11-27T15:18:37.583Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:14:09.476 15:18:37 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:14:09.476 15:18:37 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:14:09.476 15:18:37 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 92270' 00:14:09.476 15:18:37 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@973 -- # kill 92270 00:14:09.476 [2024-11-27 15:18:37.401742] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:09.476 15:18:37 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@978 -- # wait 92270 00:14:09.476 [2024-11-27 15:18:37.443126] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:14:09.736 ************************************ 00:14:09.736 END TEST raid5f_rebuild_test 00:14:09.736 ************************************ 00:14:09.736 00:14:09.736 real 0m13.709s 00:14:09.736 user 0m17.142s 00:14:09.736 sys 0m2.040s 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.736 15:18:37 bdev_raid -- bdev/bdev_raid.sh@991 -- # run_test raid5f_rebuild_test_sb raid_rebuild_test raid5f 3 true false true 00:14:09.736 15:18:37 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:14:09.736 15:18:37 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:09.736 15:18:37 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:09.736 ************************************ 00:14:09.736 START TEST raid5f_rebuild_test_sb 00:14:09.736 ************************************ 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid5f 3 true false true 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=3 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=92699 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 92699 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@835 -- # '[' -z 92699 ']' 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:09.736 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:09.736 15:18:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:09.736 [2024-11-27 15:18:37.818682] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:14:09.736 [2024-11-27 15:18:37.818921] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid92699 ] 00:14:09.736 I/O size of 3145728 is greater than zero copy threshold (65536). 00:14:09.736 Zero copy mechanism will not be used. 00:14:09.996 [2024-11-27 15:18:37.986360] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:09.996 [2024-11-27 15:18:38.013484] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:09.996 [2024-11-27 15:18:38.056204] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:09.996 [2024-11-27 15:18:38.056321] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:10.566 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:10.566 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # return 0 00:14:10.566 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:10.566 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:14:10.566 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:10.566 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:10.566 BaseBdev1_malloc 00:14:10.566 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:10.566 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:14:10.566 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:10.566 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:10.566 [2024-11-27 15:18:38.660443] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:14:10.566 [2024-11-27 15:18:38.660503] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:10.566 [2024-11-27 15:18:38.660539] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:14:10.566 [2024-11-27 15:18:38.660552] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:10.566 [2024-11-27 15:18:38.662750] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:10.566 [2024-11-27 15:18:38.662788] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:10.566 BaseBdev1 00:14:10.566 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:10.566 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:10.566 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:14:10.566 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:10.566 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:10.826 BaseBdev2_malloc 00:14:10.826 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:10.826 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:14:10.826 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:10.826 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:10.826 [2024-11-27 15:18:38.689456] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:14:10.826 [2024-11-27 15:18:38.689561] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:10.826 [2024-11-27 15:18:38.689603] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:14:10.826 [2024-11-27 15:18:38.689632] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:10.826 [2024-11-27 15:18:38.691789] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:10.826 [2024-11-27 15:18:38.691867] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:14:10.826 BaseBdev2 00:14:10.826 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:10.826 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:10.826 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:14:10.826 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:10.826 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:10.826 BaseBdev3_malloc 00:14:10.826 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:10.826 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:14:10.826 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:10.826 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:10.826 [2024-11-27 15:18:38.722484] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:14:10.826 [2024-11-27 15:18:38.722594] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:10.826 [2024-11-27 15:18:38.722638] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:14:10.826 [2024-11-27 15:18:38.722677] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:10.826 [2024-11-27 15:18:38.724912] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:10.826 [2024-11-27 15:18:38.724998] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:14:10.826 BaseBdev3 00:14:10.826 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:10.827 spare_malloc 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:10.827 spare_delay 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:10.827 [2024-11-27 15:18:38.770650] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:10.827 [2024-11-27 15:18:38.770747] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:10.827 [2024-11-27 15:18:38.770795] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:14:10.827 [2024-11-27 15:18:38.770824] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:10.827 [2024-11-27 15:18:38.773014] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:10.827 [2024-11-27 15:18:38.773085] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:10.827 spare 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:10.827 [2024-11-27 15:18:38.782694] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:10.827 [2024-11-27 15:18:38.784572] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:10.827 [2024-11-27 15:18:38.784673] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:10.827 [2024-11-27 15:18:38.784885] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:14:10.827 [2024-11-27 15:18:38.784971] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:14:10.827 [2024-11-27 15:18:38.785254] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:14:10.827 [2024-11-27 15:18:38.785721] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:14:10.827 [2024-11-27 15:18:38.785770] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:14:10.827 [2024-11-27 15:18:38.785952] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:10.827 "name": "raid_bdev1", 00:14:10.827 "uuid": "a61e22e7-3726-450a-814a-3ab52053bb1c", 00:14:10.827 "strip_size_kb": 64, 00:14:10.827 "state": "online", 00:14:10.827 "raid_level": "raid5f", 00:14:10.827 "superblock": true, 00:14:10.827 "num_base_bdevs": 3, 00:14:10.827 "num_base_bdevs_discovered": 3, 00:14:10.827 "num_base_bdevs_operational": 3, 00:14:10.827 "base_bdevs_list": [ 00:14:10.827 { 00:14:10.827 "name": "BaseBdev1", 00:14:10.827 "uuid": "f55468a8-9f2b-5b17-8f85-89353e0c0fe6", 00:14:10.827 "is_configured": true, 00:14:10.827 "data_offset": 2048, 00:14:10.827 "data_size": 63488 00:14:10.827 }, 00:14:10.827 { 00:14:10.827 "name": "BaseBdev2", 00:14:10.827 "uuid": "975e83b2-3817-523c-8d96-fadd5982b42a", 00:14:10.827 "is_configured": true, 00:14:10.827 "data_offset": 2048, 00:14:10.827 "data_size": 63488 00:14:10.827 }, 00:14:10.827 { 00:14:10.827 "name": "BaseBdev3", 00:14:10.827 "uuid": "f9ce5ce7-36ee-5b89-b1f2-3230dcd9b374", 00:14:10.827 "is_configured": true, 00:14:10.827 "data_offset": 2048, 00:14:10.827 "data_size": 63488 00:14:10.827 } 00:14:10.827 ] 00:14:10.827 }' 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:10.827 15:18:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:11.395 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:11.395 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:14:11.395 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:11.395 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:11.395 [2024-11-27 15:18:39.262765] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:11.395 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:11.395 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=126976 00:14:11.395 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:11.395 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:14:11.395 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:11.395 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:11.395 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:11.395 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:14:11.395 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:14:11.395 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:14:11.395 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:14:11.395 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:14:11.395 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:11.395 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:14:11.395 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:11.395 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:14:11.395 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:11.395 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:14:11.395 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:11.395 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:11.395 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:14:11.655 [2024-11-27 15:18:39.546208] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:14:11.655 /dev/nbd0 00:14:11.655 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:11.655 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:11.655 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:14:11.655 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:14:11.655 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:11.655 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:11.655 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:14:11.655 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:14:11.655 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:11.655 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:11.655 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:11.655 1+0 records in 00:14:11.655 1+0 records out 00:14:11.655 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000563995 s, 7.3 MB/s 00:14:11.655 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:11.655 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:14:11.655 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:11.655 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:11.655 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:14:11.655 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:11.655 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:11.655 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:14:11.655 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@630 -- # write_unit_size=256 00:14:11.655 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@631 -- # echo 128 00:14:11.655 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=131072 count=496 oflag=direct 00:14:11.916 496+0 records in 00:14:11.916 496+0 records out 00:14:11.916 65011712 bytes (65 MB, 62 MiB) copied, 0.285085 s, 228 MB/s 00:14:11.916 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:14:11.916 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:11.916 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:14:11.916 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:11.916 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:14:11.916 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:11.916 15:18:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:12.176 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:12.176 [2024-11-27 15:18:40.134150] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:12.176 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:12.176 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:12.176 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:12.176 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:12.176 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:12.176 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:14:12.176 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:14:12.176 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:14:12.176 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:12.176 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:12.177 [2024-11-27 15:18:40.154231] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:12.177 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:12.177 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:12.177 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:12.177 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:12.177 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:12.177 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:12.177 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:12.177 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:12.177 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:12.177 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:12.177 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:12.177 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:12.177 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:12.177 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:12.177 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:12.177 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:12.177 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:12.177 "name": "raid_bdev1", 00:14:12.177 "uuid": "a61e22e7-3726-450a-814a-3ab52053bb1c", 00:14:12.177 "strip_size_kb": 64, 00:14:12.177 "state": "online", 00:14:12.177 "raid_level": "raid5f", 00:14:12.177 "superblock": true, 00:14:12.177 "num_base_bdevs": 3, 00:14:12.177 "num_base_bdevs_discovered": 2, 00:14:12.177 "num_base_bdevs_operational": 2, 00:14:12.177 "base_bdevs_list": [ 00:14:12.177 { 00:14:12.177 "name": null, 00:14:12.177 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:12.177 "is_configured": false, 00:14:12.177 "data_offset": 0, 00:14:12.177 "data_size": 63488 00:14:12.177 }, 00:14:12.177 { 00:14:12.177 "name": "BaseBdev2", 00:14:12.177 "uuid": "975e83b2-3817-523c-8d96-fadd5982b42a", 00:14:12.177 "is_configured": true, 00:14:12.177 "data_offset": 2048, 00:14:12.177 "data_size": 63488 00:14:12.177 }, 00:14:12.177 { 00:14:12.177 "name": "BaseBdev3", 00:14:12.177 "uuid": "f9ce5ce7-36ee-5b89-b1f2-3230dcd9b374", 00:14:12.177 "is_configured": true, 00:14:12.177 "data_offset": 2048, 00:14:12.177 "data_size": 63488 00:14:12.177 } 00:14:12.177 ] 00:14:12.177 }' 00:14:12.177 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:12.177 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:12.744 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:12.744 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:12.744 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:12.744 [2024-11-27 15:18:40.625427] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:12.744 [2024-11-27 15:18:40.630295] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000028de0 00:14:12.744 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:12.744 15:18:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:14:12.744 [2024-11-27 15:18:40.632554] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:13.683 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:13.683 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:13.683 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:13.683 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:13.683 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:13.683 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:13.683 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:13.683 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:13.683 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:13.683 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:13.683 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:13.683 "name": "raid_bdev1", 00:14:13.683 "uuid": "a61e22e7-3726-450a-814a-3ab52053bb1c", 00:14:13.683 "strip_size_kb": 64, 00:14:13.683 "state": "online", 00:14:13.683 "raid_level": "raid5f", 00:14:13.683 "superblock": true, 00:14:13.683 "num_base_bdevs": 3, 00:14:13.683 "num_base_bdevs_discovered": 3, 00:14:13.683 "num_base_bdevs_operational": 3, 00:14:13.683 "process": { 00:14:13.683 "type": "rebuild", 00:14:13.683 "target": "spare", 00:14:13.683 "progress": { 00:14:13.683 "blocks": 20480, 00:14:13.683 "percent": 16 00:14:13.683 } 00:14:13.683 }, 00:14:13.683 "base_bdevs_list": [ 00:14:13.683 { 00:14:13.683 "name": "spare", 00:14:13.683 "uuid": "5019117e-02df-5d13-be12-43c8b4972e15", 00:14:13.683 "is_configured": true, 00:14:13.683 "data_offset": 2048, 00:14:13.683 "data_size": 63488 00:14:13.683 }, 00:14:13.683 { 00:14:13.683 "name": "BaseBdev2", 00:14:13.683 "uuid": "975e83b2-3817-523c-8d96-fadd5982b42a", 00:14:13.683 "is_configured": true, 00:14:13.683 "data_offset": 2048, 00:14:13.683 "data_size": 63488 00:14:13.683 }, 00:14:13.683 { 00:14:13.683 "name": "BaseBdev3", 00:14:13.683 "uuid": "f9ce5ce7-36ee-5b89-b1f2-3230dcd9b374", 00:14:13.683 "is_configured": true, 00:14:13.683 "data_offset": 2048, 00:14:13.683 "data_size": 63488 00:14:13.683 } 00:14:13.683 ] 00:14:13.683 }' 00:14:13.684 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:13.684 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:13.684 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:13.684 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:13.684 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:14:13.684 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:13.684 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:13.684 [2024-11-27 15:18:41.781079] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:13.945 [2024-11-27 15:18:41.841218] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:13.945 [2024-11-27 15:18:41.841293] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:13.945 [2024-11-27 15:18:41.841311] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:13.945 [2024-11-27 15:18:41.841321] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:13.945 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:13.945 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:13.945 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:13.945 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:13.945 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:13.945 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:13.945 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:13.945 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:13.945 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:13.945 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:13.945 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:13.945 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:13.945 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:13.945 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:13.945 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:13.945 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:13.945 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:13.945 "name": "raid_bdev1", 00:14:13.945 "uuid": "a61e22e7-3726-450a-814a-3ab52053bb1c", 00:14:13.945 "strip_size_kb": 64, 00:14:13.945 "state": "online", 00:14:13.945 "raid_level": "raid5f", 00:14:13.945 "superblock": true, 00:14:13.945 "num_base_bdevs": 3, 00:14:13.945 "num_base_bdevs_discovered": 2, 00:14:13.945 "num_base_bdevs_operational": 2, 00:14:13.945 "base_bdevs_list": [ 00:14:13.945 { 00:14:13.945 "name": null, 00:14:13.945 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:13.945 "is_configured": false, 00:14:13.945 "data_offset": 0, 00:14:13.945 "data_size": 63488 00:14:13.945 }, 00:14:13.945 { 00:14:13.945 "name": "BaseBdev2", 00:14:13.945 "uuid": "975e83b2-3817-523c-8d96-fadd5982b42a", 00:14:13.945 "is_configured": true, 00:14:13.945 "data_offset": 2048, 00:14:13.945 "data_size": 63488 00:14:13.945 }, 00:14:13.945 { 00:14:13.945 "name": "BaseBdev3", 00:14:13.945 "uuid": "f9ce5ce7-36ee-5b89-b1f2-3230dcd9b374", 00:14:13.945 "is_configured": true, 00:14:13.945 "data_offset": 2048, 00:14:13.945 "data_size": 63488 00:14:13.945 } 00:14:13.945 ] 00:14:13.945 }' 00:14:13.945 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:13.945 15:18:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:14.515 15:18:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:14.515 15:18:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:14.515 15:18:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:14.515 15:18:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:14.515 15:18:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:14.515 15:18:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:14.515 15:18:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:14.515 15:18:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:14.515 15:18:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:14.515 15:18:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:14.515 15:18:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:14.515 "name": "raid_bdev1", 00:14:14.515 "uuid": "a61e22e7-3726-450a-814a-3ab52053bb1c", 00:14:14.515 "strip_size_kb": 64, 00:14:14.515 "state": "online", 00:14:14.515 "raid_level": "raid5f", 00:14:14.515 "superblock": true, 00:14:14.515 "num_base_bdevs": 3, 00:14:14.515 "num_base_bdevs_discovered": 2, 00:14:14.515 "num_base_bdevs_operational": 2, 00:14:14.515 "base_bdevs_list": [ 00:14:14.515 { 00:14:14.515 "name": null, 00:14:14.515 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:14.515 "is_configured": false, 00:14:14.515 "data_offset": 0, 00:14:14.515 "data_size": 63488 00:14:14.515 }, 00:14:14.515 { 00:14:14.515 "name": "BaseBdev2", 00:14:14.515 "uuid": "975e83b2-3817-523c-8d96-fadd5982b42a", 00:14:14.515 "is_configured": true, 00:14:14.515 "data_offset": 2048, 00:14:14.515 "data_size": 63488 00:14:14.515 }, 00:14:14.515 { 00:14:14.515 "name": "BaseBdev3", 00:14:14.515 "uuid": "f9ce5ce7-36ee-5b89-b1f2-3230dcd9b374", 00:14:14.515 "is_configured": true, 00:14:14.515 "data_offset": 2048, 00:14:14.515 "data_size": 63488 00:14:14.515 } 00:14:14.515 ] 00:14:14.515 }' 00:14:14.515 15:18:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:14.515 15:18:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:14.515 15:18:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:14.515 15:18:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:14.515 15:18:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:14.515 15:18:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:14.515 15:18:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:14.515 [2024-11-27 15:18:42.478689] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:14.515 [2024-11-27 15:18:42.483382] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000028eb0 00:14:14.515 15:18:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:14.515 15:18:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:14:14.515 [2024-11-27 15:18:42.485529] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:15.456 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:15.456 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:15.456 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:15.456 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:15.456 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:15.456 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:15.456 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:15.456 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:15.456 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:15.456 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:15.456 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:15.456 "name": "raid_bdev1", 00:14:15.456 "uuid": "a61e22e7-3726-450a-814a-3ab52053bb1c", 00:14:15.456 "strip_size_kb": 64, 00:14:15.456 "state": "online", 00:14:15.456 "raid_level": "raid5f", 00:14:15.456 "superblock": true, 00:14:15.456 "num_base_bdevs": 3, 00:14:15.456 "num_base_bdevs_discovered": 3, 00:14:15.456 "num_base_bdevs_operational": 3, 00:14:15.456 "process": { 00:14:15.456 "type": "rebuild", 00:14:15.456 "target": "spare", 00:14:15.456 "progress": { 00:14:15.456 "blocks": 20480, 00:14:15.456 "percent": 16 00:14:15.456 } 00:14:15.456 }, 00:14:15.456 "base_bdevs_list": [ 00:14:15.456 { 00:14:15.456 "name": "spare", 00:14:15.456 "uuid": "5019117e-02df-5d13-be12-43c8b4972e15", 00:14:15.456 "is_configured": true, 00:14:15.456 "data_offset": 2048, 00:14:15.456 "data_size": 63488 00:14:15.456 }, 00:14:15.456 { 00:14:15.456 "name": "BaseBdev2", 00:14:15.456 "uuid": "975e83b2-3817-523c-8d96-fadd5982b42a", 00:14:15.456 "is_configured": true, 00:14:15.456 "data_offset": 2048, 00:14:15.456 "data_size": 63488 00:14:15.456 }, 00:14:15.456 { 00:14:15.456 "name": "BaseBdev3", 00:14:15.456 "uuid": "f9ce5ce7-36ee-5b89-b1f2-3230dcd9b374", 00:14:15.456 "is_configured": true, 00:14:15.456 "data_offset": 2048, 00:14:15.456 "data_size": 63488 00:14:15.456 } 00:14:15.456 ] 00:14:15.456 }' 00:14:15.456 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:15.715 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:15.715 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:15.715 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:15.715 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:14:15.715 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:14:15.715 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:14:15.715 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=3 00:14:15.715 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:14:15.715 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=464 00:14:15.715 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:15.715 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:15.715 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:15.715 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:15.715 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:15.715 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:15.715 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:15.715 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:15.715 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:15.715 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:15.716 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:15.716 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:15.716 "name": "raid_bdev1", 00:14:15.716 "uuid": "a61e22e7-3726-450a-814a-3ab52053bb1c", 00:14:15.716 "strip_size_kb": 64, 00:14:15.716 "state": "online", 00:14:15.716 "raid_level": "raid5f", 00:14:15.716 "superblock": true, 00:14:15.716 "num_base_bdevs": 3, 00:14:15.716 "num_base_bdevs_discovered": 3, 00:14:15.716 "num_base_bdevs_operational": 3, 00:14:15.716 "process": { 00:14:15.716 "type": "rebuild", 00:14:15.716 "target": "spare", 00:14:15.716 "progress": { 00:14:15.716 "blocks": 22528, 00:14:15.716 "percent": 17 00:14:15.716 } 00:14:15.716 }, 00:14:15.716 "base_bdevs_list": [ 00:14:15.716 { 00:14:15.716 "name": "spare", 00:14:15.716 "uuid": "5019117e-02df-5d13-be12-43c8b4972e15", 00:14:15.716 "is_configured": true, 00:14:15.716 "data_offset": 2048, 00:14:15.716 "data_size": 63488 00:14:15.716 }, 00:14:15.716 { 00:14:15.716 "name": "BaseBdev2", 00:14:15.716 "uuid": "975e83b2-3817-523c-8d96-fadd5982b42a", 00:14:15.716 "is_configured": true, 00:14:15.716 "data_offset": 2048, 00:14:15.716 "data_size": 63488 00:14:15.716 }, 00:14:15.716 { 00:14:15.716 "name": "BaseBdev3", 00:14:15.716 "uuid": "f9ce5ce7-36ee-5b89-b1f2-3230dcd9b374", 00:14:15.716 "is_configured": true, 00:14:15.716 "data_offset": 2048, 00:14:15.716 "data_size": 63488 00:14:15.716 } 00:14:15.716 ] 00:14:15.716 }' 00:14:15.716 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:15.716 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:15.716 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:15.716 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:15.716 15:18:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:17.111 15:18:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:17.111 15:18:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:17.111 15:18:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:17.111 15:18:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:17.111 15:18:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:17.111 15:18:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:17.111 15:18:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:17.111 15:18:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:17.111 15:18:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.111 15:18:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.112 15:18:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.112 15:18:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:17.112 "name": "raid_bdev1", 00:14:17.112 "uuid": "a61e22e7-3726-450a-814a-3ab52053bb1c", 00:14:17.112 "strip_size_kb": 64, 00:14:17.112 "state": "online", 00:14:17.112 "raid_level": "raid5f", 00:14:17.112 "superblock": true, 00:14:17.112 "num_base_bdevs": 3, 00:14:17.112 "num_base_bdevs_discovered": 3, 00:14:17.112 "num_base_bdevs_operational": 3, 00:14:17.112 "process": { 00:14:17.112 "type": "rebuild", 00:14:17.112 "target": "spare", 00:14:17.112 "progress": { 00:14:17.112 "blocks": 47104, 00:14:17.112 "percent": 37 00:14:17.112 } 00:14:17.112 }, 00:14:17.112 "base_bdevs_list": [ 00:14:17.112 { 00:14:17.112 "name": "spare", 00:14:17.112 "uuid": "5019117e-02df-5d13-be12-43c8b4972e15", 00:14:17.112 "is_configured": true, 00:14:17.112 "data_offset": 2048, 00:14:17.112 "data_size": 63488 00:14:17.112 }, 00:14:17.112 { 00:14:17.112 "name": "BaseBdev2", 00:14:17.112 "uuid": "975e83b2-3817-523c-8d96-fadd5982b42a", 00:14:17.112 "is_configured": true, 00:14:17.112 "data_offset": 2048, 00:14:17.112 "data_size": 63488 00:14:17.112 }, 00:14:17.112 { 00:14:17.112 "name": "BaseBdev3", 00:14:17.112 "uuid": "f9ce5ce7-36ee-5b89-b1f2-3230dcd9b374", 00:14:17.112 "is_configured": true, 00:14:17.112 "data_offset": 2048, 00:14:17.112 "data_size": 63488 00:14:17.112 } 00:14:17.112 ] 00:14:17.112 }' 00:14:17.112 15:18:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:17.112 15:18:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:17.112 15:18:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:17.112 15:18:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:17.112 15:18:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:18.048 15:18:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:18.048 15:18:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:18.048 15:18:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:18.048 15:18:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:18.048 15:18:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:18.048 15:18:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:18.048 15:18:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:18.048 15:18:45 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:18.048 15:18:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:18.048 15:18:45 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:18.048 15:18:45 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:18.048 15:18:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:18.048 "name": "raid_bdev1", 00:14:18.048 "uuid": "a61e22e7-3726-450a-814a-3ab52053bb1c", 00:14:18.048 "strip_size_kb": 64, 00:14:18.048 "state": "online", 00:14:18.048 "raid_level": "raid5f", 00:14:18.048 "superblock": true, 00:14:18.048 "num_base_bdevs": 3, 00:14:18.048 "num_base_bdevs_discovered": 3, 00:14:18.048 "num_base_bdevs_operational": 3, 00:14:18.048 "process": { 00:14:18.048 "type": "rebuild", 00:14:18.048 "target": "spare", 00:14:18.048 "progress": { 00:14:18.048 "blocks": 69632, 00:14:18.048 "percent": 54 00:14:18.048 } 00:14:18.048 }, 00:14:18.048 "base_bdevs_list": [ 00:14:18.048 { 00:14:18.048 "name": "spare", 00:14:18.048 "uuid": "5019117e-02df-5d13-be12-43c8b4972e15", 00:14:18.048 "is_configured": true, 00:14:18.048 "data_offset": 2048, 00:14:18.048 "data_size": 63488 00:14:18.048 }, 00:14:18.048 { 00:14:18.048 "name": "BaseBdev2", 00:14:18.048 "uuid": "975e83b2-3817-523c-8d96-fadd5982b42a", 00:14:18.048 "is_configured": true, 00:14:18.048 "data_offset": 2048, 00:14:18.048 "data_size": 63488 00:14:18.048 }, 00:14:18.049 { 00:14:18.049 "name": "BaseBdev3", 00:14:18.049 "uuid": "f9ce5ce7-36ee-5b89-b1f2-3230dcd9b374", 00:14:18.049 "is_configured": true, 00:14:18.049 "data_offset": 2048, 00:14:18.049 "data_size": 63488 00:14:18.049 } 00:14:18.049 ] 00:14:18.049 }' 00:14:18.049 15:18:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:18.049 15:18:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:18.049 15:18:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:18.049 15:18:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:18.049 15:18:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:19.429 15:18:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:19.429 15:18:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:19.429 15:18:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:19.429 15:18:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:19.429 15:18:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:19.429 15:18:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:19.429 15:18:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:19.429 15:18:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:19.429 15:18:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:19.429 15:18:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:19.429 15:18:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:19.429 15:18:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:19.429 "name": "raid_bdev1", 00:14:19.429 "uuid": "a61e22e7-3726-450a-814a-3ab52053bb1c", 00:14:19.429 "strip_size_kb": 64, 00:14:19.429 "state": "online", 00:14:19.429 "raid_level": "raid5f", 00:14:19.429 "superblock": true, 00:14:19.429 "num_base_bdevs": 3, 00:14:19.429 "num_base_bdevs_discovered": 3, 00:14:19.429 "num_base_bdevs_operational": 3, 00:14:19.429 "process": { 00:14:19.429 "type": "rebuild", 00:14:19.429 "target": "spare", 00:14:19.429 "progress": { 00:14:19.429 "blocks": 92160, 00:14:19.429 "percent": 72 00:14:19.429 } 00:14:19.429 }, 00:14:19.429 "base_bdevs_list": [ 00:14:19.429 { 00:14:19.429 "name": "spare", 00:14:19.429 "uuid": "5019117e-02df-5d13-be12-43c8b4972e15", 00:14:19.429 "is_configured": true, 00:14:19.429 "data_offset": 2048, 00:14:19.429 "data_size": 63488 00:14:19.429 }, 00:14:19.429 { 00:14:19.429 "name": "BaseBdev2", 00:14:19.429 "uuid": "975e83b2-3817-523c-8d96-fadd5982b42a", 00:14:19.429 "is_configured": true, 00:14:19.429 "data_offset": 2048, 00:14:19.429 "data_size": 63488 00:14:19.429 }, 00:14:19.429 { 00:14:19.429 "name": "BaseBdev3", 00:14:19.429 "uuid": "f9ce5ce7-36ee-5b89-b1f2-3230dcd9b374", 00:14:19.429 "is_configured": true, 00:14:19.429 "data_offset": 2048, 00:14:19.429 "data_size": 63488 00:14:19.429 } 00:14:19.429 ] 00:14:19.429 }' 00:14:19.429 15:18:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:19.429 15:18:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:19.429 15:18:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:19.429 15:18:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:19.429 15:18:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:20.368 15:18:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:20.368 15:18:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:20.368 15:18:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:20.368 15:18:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:20.368 15:18:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:20.368 15:18:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:20.368 15:18:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:20.368 15:18:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:20.368 15:18:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:20.368 15:18:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:20.368 15:18:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:20.368 15:18:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:20.368 "name": "raid_bdev1", 00:14:20.368 "uuid": "a61e22e7-3726-450a-814a-3ab52053bb1c", 00:14:20.368 "strip_size_kb": 64, 00:14:20.368 "state": "online", 00:14:20.368 "raid_level": "raid5f", 00:14:20.368 "superblock": true, 00:14:20.368 "num_base_bdevs": 3, 00:14:20.368 "num_base_bdevs_discovered": 3, 00:14:20.368 "num_base_bdevs_operational": 3, 00:14:20.368 "process": { 00:14:20.368 "type": "rebuild", 00:14:20.368 "target": "spare", 00:14:20.368 "progress": { 00:14:20.368 "blocks": 116736, 00:14:20.368 "percent": 91 00:14:20.368 } 00:14:20.368 }, 00:14:20.368 "base_bdevs_list": [ 00:14:20.368 { 00:14:20.368 "name": "spare", 00:14:20.368 "uuid": "5019117e-02df-5d13-be12-43c8b4972e15", 00:14:20.368 "is_configured": true, 00:14:20.368 "data_offset": 2048, 00:14:20.368 "data_size": 63488 00:14:20.368 }, 00:14:20.368 { 00:14:20.368 "name": "BaseBdev2", 00:14:20.368 "uuid": "975e83b2-3817-523c-8d96-fadd5982b42a", 00:14:20.368 "is_configured": true, 00:14:20.368 "data_offset": 2048, 00:14:20.368 "data_size": 63488 00:14:20.368 }, 00:14:20.368 { 00:14:20.368 "name": "BaseBdev3", 00:14:20.368 "uuid": "f9ce5ce7-36ee-5b89-b1f2-3230dcd9b374", 00:14:20.368 "is_configured": true, 00:14:20.368 "data_offset": 2048, 00:14:20.368 "data_size": 63488 00:14:20.369 } 00:14:20.369 ] 00:14:20.369 }' 00:14:20.369 15:18:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:20.369 15:18:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:20.369 15:18:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:20.369 15:18:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:20.369 15:18:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:20.628 [2024-11-27 15:18:48.725198] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:14:20.628 [2024-11-27 15:18:48.725263] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:14:20.628 [2024-11-27 15:18:48.725373] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:21.567 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:21.567 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:21.567 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:21.567 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:21.567 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:21.567 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:21.567 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:21.567 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:21.567 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:21.567 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:21.567 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:21.567 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:21.567 "name": "raid_bdev1", 00:14:21.567 "uuid": "a61e22e7-3726-450a-814a-3ab52053bb1c", 00:14:21.567 "strip_size_kb": 64, 00:14:21.567 "state": "online", 00:14:21.567 "raid_level": "raid5f", 00:14:21.567 "superblock": true, 00:14:21.567 "num_base_bdevs": 3, 00:14:21.567 "num_base_bdevs_discovered": 3, 00:14:21.567 "num_base_bdevs_operational": 3, 00:14:21.567 "base_bdevs_list": [ 00:14:21.567 { 00:14:21.567 "name": "spare", 00:14:21.567 "uuid": "5019117e-02df-5d13-be12-43c8b4972e15", 00:14:21.567 "is_configured": true, 00:14:21.567 "data_offset": 2048, 00:14:21.567 "data_size": 63488 00:14:21.567 }, 00:14:21.567 { 00:14:21.567 "name": "BaseBdev2", 00:14:21.568 "uuid": "975e83b2-3817-523c-8d96-fadd5982b42a", 00:14:21.568 "is_configured": true, 00:14:21.568 "data_offset": 2048, 00:14:21.568 "data_size": 63488 00:14:21.568 }, 00:14:21.568 { 00:14:21.568 "name": "BaseBdev3", 00:14:21.568 "uuid": "f9ce5ce7-36ee-5b89-b1f2-3230dcd9b374", 00:14:21.568 "is_configured": true, 00:14:21.568 "data_offset": 2048, 00:14:21.568 "data_size": 63488 00:14:21.568 } 00:14:21.568 ] 00:14:21.568 }' 00:14:21.568 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:21.568 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:14:21.568 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:21.568 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:14:21.568 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:14:21.568 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:21.568 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:21.568 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:21.568 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:21.568 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:21.568 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:21.568 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:21.568 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:21.568 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:21.568 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:21.568 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:21.568 "name": "raid_bdev1", 00:14:21.568 "uuid": "a61e22e7-3726-450a-814a-3ab52053bb1c", 00:14:21.568 "strip_size_kb": 64, 00:14:21.568 "state": "online", 00:14:21.568 "raid_level": "raid5f", 00:14:21.568 "superblock": true, 00:14:21.568 "num_base_bdevs": 3, 00:14:21.568 "num_base_bdevs_discovered": 3, 00:14:21.568 "num_base_bdevs_operational": 3, 00:14:21.568 "base_bdevs_list": [ 00:14:21.568 { 00:14:21.568 "name": "spare", 00:14:21.568 "uuid": "5019117e-02df-5d13-be12-43c8b4972e15", 00:14:21.568 "is_configured": true, 00:14:21.568 "data_offset": 2048, 00:14:21.568 "data_size": 63488 00:14:21.568 }, 00:14:21.568 { 00:14:21.568 "name": "BaseBdev2", 00:14:21.568 "uuid": "975e83b2-3817-523c-8d96-fadd5982b42a", 00:14:21.568 "is_configured": true, 00:14:21.568 "data_offset": 2048, 00:14:21.568 "data_size": 63488 00:14:21.568 }, 00:14:21.568 { 00:14:21.568 "name": "BaseBdev3", 00:14:21.568 "uuid": "f9ce5ce7-36ee-5b89-b1f2-3230dcd9b374", 00:14:21.568 "is_configured": true, 00:14:21.568 "data_offset": 2048, 00:14:21.568 "data_size": 63488 00:14:21.568 } 00:14:21.568 ] 00:14:21.568 }' 00:14:21.568 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:21.568 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:21.568 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:21.827 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:21.827 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:21.827 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:21.827 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:21.827 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:21.827 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:21.827 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:21.827 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:21.827 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:21.827 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:21.827 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:21.827 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:21.827 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:21.827 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:21.827 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:21.827 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:21.827 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:21.827 "name": "raid_bdev1", 00:14:21.827 "uuid": "a61e22e7-3726-450a-814a-3ab52053bb1c", 00:14:21.827 "strip_size_kb": 64, 00:14:21.827 "state": "online", 00:14:21.827 "raid_level": "raid5f", 00:14:21.827 "superblock": true, 00:14:21.827 "num_base_bdevs": 3, 00:14:21.827 "num_base_bdevs_discovered": 3, 00:14:21.827 "num_base_bdevs_operational": 3, 00:14:21.827 "base_bdevs_list": [ 00:14:21.827 { 00:14:21.827 "name": "spare", 00:14:21.827 "uuid": "5019117e-02df-5d13-be12-43c8b4972e15", 00:14:21.827 "is_configured": true, 00:14:21.827 "data_offset": 2048, 00:14:21.827 "data_size": 63488 00:14:21.827 }, 00:14:21.827 { 00:14:21.827 "name": "BaseBdev2", 00:14:21.827 "uuid": "975e83b2-3817-523c-8d96-fadd5982b42a", 00:14:21.827 "is_configured": true, 00:14:21.827 "data_offset": 2048, 00:14:21.827 "data_size": 63488 00:14:21.827 }, 00:14:21.827 { 00:14:21.827 "name": "BaseBdev3", 00:14:21.827 "uuid": "f9ce5ce7-36ee-5b89-b1f2-3230dcd9b374", 00:14:21.827 "is_configured": true, 00:14:21.827 "data_offset": 2048, 00:14:21.827 "data_size": 63488 00:14:21.827 } 00:14:21.827 ] 00:14:21.827 }' 00:14:21.827 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:21.827 15:18:49 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.086 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:22.086 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:22.086 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.086 [2024-11-27 15:18:50.140813] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:22.086 [2024-11-27 15:18:50.140949] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:22.086 [2024-11-27 15:18:50.141093] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:22.086 [2024-11-27 15:18:50.141236] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:22.086 [2024-11-27 15:18:50.141297] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:14:22.086 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:22.086 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:22.086 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:14:22.086 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:22.086 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.086 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:22.346 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:14:22.346 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:14:22.346 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:14:22.346 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:14:22.346 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:22.346 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:14:22.346 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:22.346 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:14:22.346 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:22.346 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:14:22.346 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:22.346 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:22.346 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:14:22.346 /dev/nbd0 00:14:22.346 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:22.346 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:22.346 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:14:22.346 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:14:22.346 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:22.346 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:22.346 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:14:22.346 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:14:22.346 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:22.346 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:22.346 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:22.346 1+0 records in 00:14:22.346 1+0 records out 00:14:22.346 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00054003 s, 7.6 MB/s 00:14:22.346 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:22.346 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:14:22.346 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:22.606 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:22.606 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:14:22.606 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:22.606 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:22.606 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:14:22.606 /dev/nbd1 00:14:22.606 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:14:22.606 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:14:22.606 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:14:22.606 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:14:22.606 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:22.606 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:22.606 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:14:22.606 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:14:22.606 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:22.606 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:22.606 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:22.606 1+0 records in 00:14:22.606 1+0 records out 00:14:22.606 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000409284 s, 10.0 MB/s 00:14:22.606 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:22.606 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:14:22.606 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:22.606 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:22.606 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:14:22.606 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:22.606 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:22.606 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:14:22.865 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:14:22.865 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:22.865 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:14:22.865 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:22.865 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:14:22.865 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:22.865 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:22.865 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:23.125 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:23.125 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:23.125 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:23.125 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:23.125 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:23.125 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:14:23.125 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:14:23.125 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:23.125 15:18:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:14:23.125 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:14:23.125 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:14:23.125 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:14:23.125 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:23.125 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:23.125 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:14:23.125 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:14:23.125 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:14:23.125 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:14:23.125 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:14:23.125 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:23.125 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:23.125 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:23.125 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:23.125 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:23.125 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:23.125 [2024-11-27 15:18:51.203006] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:23.125 [2024-11-27 15:18:51.203067] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:23.125 [2024-11-27 15:18:51.203090] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:14:23.125 [2024-11-27 15:18:51.203098] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:23.125 [2024-11-27 15:18:51.205326] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:23.125 [2024-11-27 15:18:51.205366] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:23.125 [2024-11-27 15:18:51.205456] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:14:23.125 [2024-11-27 15:18:51.205505] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:23.125 [2024-11-27 15:18:51.205623] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:23.125 [2024-11-27 15:18:51.205714] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:23.125 spare 00:14:23.125 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:23.125 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:14:23.125 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:23.125 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:23.385 [2024-11-27 15:18:51.305604] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006600 00:14:23.385 [2024-11-27 15:18:51.305630] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:14:23.385 [2024-11-27 15:18:51.305867] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000047560 00:14:23.385 [2024-11-27 15:18:51.306298] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006600 00:14:23.385 [2024-11-27 15:18:51.306318] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006600 00:14:23.385 [2024-11-27 15:18:51.306465] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:23.385 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:23.385 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:23.385 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:23.385 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:23.385 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:23.385 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:23.385 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:23.385 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:23.385 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:23.385 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:23.385 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:23.385 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:23.385 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:23.385 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:23.385 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:23.385 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:23.385 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:23.385 "name": "raid_bdev1", 00:14:23.385 "uuid": "a61e22e7-3726-450a-814a-3ab52053bb1c", 00:14:23.385 "strip_size_kb": 64, 00:14:23.385 "state": "online", 00:14:23.385 "raid_level": "raid5f", 00:14:23.385 "superblock": true, 00:14:23.385 "num_base_bdevs": 3, 00:14:23.385 "num_base_bdevs_discovered": 3, 00:14:23.385 "num_base_bdevs_operational": 3, 00:14:23.385 "base_bdevs_list": [ 00:14:23.385 { 00:14:23.385 "name": "spare", 00:14:23.385 "uuid": "5019117e-02df-5d13-be12-43c8b4972e15", 00:14:23.385 "is_configured": true, 00:14:23.385 "data_offset": 2048, 00:14:23.385 "data_size": 63488 00:14:23.385 }, 00:14:23.385 { 00:14:23.385 "name": "BaseBdev2", 00:14:23.385 "uuid": "975e83b2-3817-523c-8d96-fadd5982b42a", 00:14:23.385 "is_configured": true, 00:14:23.385 "data_offset": 2048, 00:14:23.385 "data_size": 63488 00:14:23.385 }, 00:14:23.385 { 00:14:23.385 "name": "BaseBdev3", 00:14:23.385 "uuid": "f9ce5ce7-36ee-5b89-b1f2-3230dcd9b374", 00:14:23.385 "is_configured": true, 00:14:23.385 "data_offset": 2048, 00:14:23.385 "data_size": 63488 00:14:23.385 } 00:14:23.385 ] 00:14:23.385 }' 00:14:23.385 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:23.385 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:23.955 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:23.955 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:23.955 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:23.955 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:23.955 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:23.955 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:23.955 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:23.955 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:23.955 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:23.955 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:23.955 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:23.955 "name": "raid_bdev1", 00:14:23.955 "uuid": "a61e22e7-3726-450a-814a-3ab52053bb1c", 00:14:23.955 "strip_size_kb": 64, 00:14:23.955 "state": "online", 00:14:23.955 "raid_level": "raid5f", 00:14:23.955 "superblock": true, 00:14:23.955 "num_base_bdevs": 3, 00:14:23.955 "num_base_bdevs_discovered": 3, 00:14:23.955 "num_base_bdevs_operational": 3, 00:14:23.955 "base_bdevs_list": [ 00:14:23.955 { 00:14:23.955 "name": "spare", 00:14:23.955 "uuid": "5019117e-02df-5d13-be12-43c8b4972e15", 00:14:23.955 "is_configured": true, 00:14:23.955 "data_offset": 2048, 00:14:23.955 "data_size": 63488 00:14:23.955 }, 00:14:23.955 { 00:14:23.955 "name": "BaseBdev2", 00:14:23.955 "uuid": "975e83b2-3817-523c-8d96-fadd5982b42a", 00:14:23.955 "is_configured": true, 00:14:23.955 "data_offset": 2048, 00:14:23.955 "data_size": 63488 00:14:23.955 }, 00:14:23.955 { 00:14:23.955 "name": "BaseBdev3", 00:14:23.955 "uuid": "f9ce5ce7-36ee-5b89-b1f2-3230dcd9b374", 00:14:23.955 "is_configured": true, 00:14:23.955 "data_offset": 2048, 00:14:23.955 "data_size": 63488 00:14:23.955 } 00:14:23.955 ] 00:14:23.955 }' 00:14:23.955 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:23.955 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:23.955 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:23.955 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:23.955 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:23.955 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:23.955 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:23.955 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:14:23.955 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:23.955 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:14:23.955 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:14:23.955 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:23.955 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:23.955 [2024-11-27 15:18:51.986674] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:23.955 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:23.955 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:23.955 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:23.955 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:23.955 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:23.956 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:23.956 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:23.956 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:23.956 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:23.956 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:23.956 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:23.956 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:23.956 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:23.956 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:23.956 15:18:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:23.956 15:18:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:23.956 15:18:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:23.956 "name": "raid_bdev1", 00:14:23.956 "uuid": "a61e22e7-3726-450a-814a-3ab52053bb1c", 00:14:23.956 "strip_size_kb": 64, 00:14:23.956 "state": "online", 00:14:23.956 "raid_level": "raid5f", 00:14:23.956 "superblock": true, 00:14:23.956 "num_base_bdevs": 3, 00:14:23.956 "num_base_bdevs_discovered": 2, 00:14:23.956 "num_base_bdevs_operational": 2, 00:14:23.956 "base_bdevs_list": [ 00:14:23.956 { 00:14:23.956 "name": null, 00:14:23.956 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:23.956 "is_configured": false, 00:14:23.956 "data_offset": 0, 00:14:23.956 "data_size": 63488 00:14:23.956 }, 00:14:23.956 { 00:14:23.956 "name": "BaseBdev2", 00:14:23.956 "uuid": "975e83b2-3817-523c-8d96-fadd5982b42a", 00:14:23.956 "is_configured": true, 00:14:23.956 "data_offset": 2048, 00:14:23.956 "data_size": 63488 00:14:23.956 }, 00:14:23.956 { 00:14:23.956 "name": "BaseBdev3", 00:14:23.956 "uuid": "f9ce5ce7-36ee-5b89-b1f2-3230dcd9b374", 00:14:23.956 "is_configured": true, 00:14:23.956 "data_offset": 2048, 00:14:23.956 "data_size": 63488 00:14:23.956 } 00:14:23.956 ] 00:14:23.956 }' 00:14:23.956 15:18:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:23.956 15:18:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.526 15:18:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:24.526 15:18:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:24.526 15:18:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.526 [2024-11-27 15:18:52.441935] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:24.526 [2024-11-27 15:18:52.442125] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:14:24.526 [2024-11-27 15:18:52.442142] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:14:24.526 [2024-11-27 15:18:52.442181] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:24.526 [2024-11-27 15:18:52.446553] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000047630 00:14:24.526 15:18:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:24.526 15:18:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:14:24.526 [2024-11-27 15:18:52.448724] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:25.464 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:25.464 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:25.464 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:25.464 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:25.464 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:25.464 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:25.465 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:25.465 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:25.465 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:25.465 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:25.465 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:25.465 "name": "raid_bdev1", 00:14:25.465 "uuid": "a61e22e7-3726-450a-814a-3ab52053bb1c", 00:14:25.465 "strip_size_kb": 64, 00:14:25.465 "state": "online", 00:14:25.465 "raid_level": "raid5f", 00:14:25.465 "superblock": true, 00:14:25.465 "num_base_bdevs": 3, 00:14:25.465 "num_base_bdevs_discovered": 3, 00:14:25.465 "num_base_bdevs_operational": 3, 00:14:25.465 "process": { 00:14:25.465 "type": "rebuild", 00:14:25.465 "target": "spare", 00:14:25.465 "progress": { 00:14:25.465 "blocks": 20480, 00:14:25.465 "percent": 16 00:14:25.465 } 00:14:25.465 }, 00:14:25.465 "base_bdevs_list": [ 00:14:25.465 { 00:14:25.465 "name": "spare", 00:14:25.465 "uuid": "5019117e-02df-5d13-be12-43c8b4972e15", 00:14:25.465 "is_configured": true, 00:14:25.465 "data_offset": 2048, 00:14:25.465 "data_size": 63488 00:14:25.465 }, 00:14:25.465 { 00:14:25.465 "name": "BaseBdev2", 00:14:25.465 "uuid": "975e83b2-3817-523c-8d96-fadd5982b42a", 00:14:25.465 "is_configured": true, 00:14:25.465 "data_offset": 2048, 00:14:25.465 "data_size": 63488 00:14:25.465 }, 00:14:25.465 { 00:14:25.465 "name": "BaseBdev3", 00:14:25.465 "uuid": "f9ce5ce7-36ee-5b89-b1f2-3230dcd9b374", 00:14:25.465 "is_configured": true, 00:14:25.465 "data_offset": 2048, 00:14:25.465 "data_size": 63488 00:14:25.465 } 00:14:25.465 ] 00:14:25.465 }' 00:14:25.465 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:25.465 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:25.465 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:25.724 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:25.724 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:14:25.724 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:25.724 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:25.725 [2024-11-27 15:18:53.612976] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:25.725 [2024-11-27 15:18:53.655927] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:25.725 [2024-11-27 15:18:53.656028] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:25.725 [2024-11-27 15:18:53.656067] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:25.725 [2024-11-27 15:18:53.656088] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:25.725 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:25.725 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:25.725 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:25.725 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:25.725 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:25.725 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:25.725 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:25.725 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:25.725 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:25.725 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:25.725 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:25.725 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:25.725 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:25.725 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:25.725 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:25.725 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:25.725 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:25.725 "name": "raid_bdev1", 00:14:25.725 "uuid": "a61e22e7-3726-450a-814a-3ab52053bb1c", 00:14:25.725 "strip_size_kb": 64, 00:14:25.725 "state": "online", 00:14:25.725 "raid_level": "raid5f", 00:14:25.725 "superblock": true, 00:14:25.725 "num_base_bdevs": 3, 00:14:25.725 "num_base_bdevs_discovered": 2, 00:14:25.725 "num_base_bdevs_operational": 2, 00:14:25.725 "base_bdevs_list": [ 00:14:25.725 { 00:14:25.725 "name": null, 00:14:25.725 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:25.725 "is_configured": false, 00:14:25.725 "data_offset": 0, 00:14:25.725 "data_size": 63488 00:14:25.725 }, 00:14:25.725 { 00:14:25.725 "name": "BaseBdev2", 00:14:25.725 "uuid": "975e83b2-3817-523c-8d96-fadd5982b42a", 00:14:25.725 "is_configured": true, 00:14:25.725 "data_offset": 2048, 00:14:25.725 "data_size": 63488 00:14:25.725 }, 00:14:25.725 { 00:14:25.725 "name": "BaseBdev3", 00:14:25.725 "uuid": "f9ce5ce7-36ee-5b89-b1f2-3230dcd9b374", 00:14:25.725 "is_configured": true, 00:14:25.725 "data_offset": 2048, 00:14:25.725 "data_size": 63488 00:14:25.725 } 00:14:25.725 ] 00:14:25.725 }' 00:14:25.725 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:25.725 15:18:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:26.292 15:18:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:26.292 15:18:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.292 15:18:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:26.292 [2024-11-27 15:18:54.152681] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:26.292 [2024-11-27 15:18:54.152801] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:26.292 [2024-11-27 15:18:54.152843] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b780 00:14:26.292 [2024-11-27 15:18:54.152872] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:26.292 [2024-11-27 15:18:54.153336] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:26.292 [2024-11-27 15:18:54.153399] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:26.292 [2024-11-27 15:18:54.153513] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:14:26.292 [2024-11-27 15:18:54.153554] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:14:26.292 [2024-11-27 15:18:54.153599] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:14:26.292 [2024-11-27 15:18:54.153641] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:26.292 [2024-11-27 15:18:54.157926] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000047700 00:14:26.292 spare 00:14:26.292 15:18:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.292 15:18:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:14:26.292 [2024-11-27 15:18:54.160132] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:27.260 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:27.260 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:27.260 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:27.260 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:27.260 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:27.260 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:27.260 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:27.260 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:27.260 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:27.260 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:27.260 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:27.260 "name": "raid_bdev1", 00:14:27.260 "uuid": "a61e22e7-3726-450a-814a-3ab52053bb1c", 00:14:27.260 "strip_size_kb": 64, 00:14:27.260 "state": "online", 00:14:27.260 "raid_level": "raid5f", 00:14:27.260 "superblock": true, 00:14:27.260 "num_base_bdevs": 3, 00:14:27.260 "num_base_bdevs_discovered": 3, 00:14:27.260 "num_base_bdevs_operational": 3, 00:14:27.260 "process": { 00:14:27.260 "type": "rebuild", 00:14:27.260 "target": "spare", 00:14:27.260 "progress": { 00:14:27.260 "blocks": 20480, 00:14:27.260 "percent": 16 00:14:27.260 } 00:14:27.260 }, 00:14:27.260 "base_bdevs_list": [ 00:14:27.260 { 00:14:27.260 "name": "spare", 00:14:27.260 "uuid": "5019117e-02df-5d13-be12-43c8b4972e15", 00:14:27.260 "is_configured": true, 00:14:27.260 "data_offset": 2048, 00:14:27.260 "data_size": 63488 00:14:27.260 }, 00:14:27.260 { 00:14:27.260 "name": "BaseBdev2", 00:14:27.260 "uuid": "975e83b2-3817-523c-8d96-fadd5982b42a", 00:14:27.260 "is_configured": true, 00:14:27.260 "data_offset": 2048, 00:14:27.260 "data_size": 63488 00:14:27.260 }, 00:14:27.260 { 00:14:27.260 "name": "BaseBdev3", 00:14:27.260 "uuid": "f9ce5ce7-36ee-5b89-b1f2-3230dcd9b374", 00:14:27.260 "is_configured": true, 00:14:27.260 "data_offset": 2048, 00:14:27.260 "data_size": 63488 00:14:27.260 } 00:14:27.260 ] 00:14:27.260 }' 00:14:27.260 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:27.260 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:27.260 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:27.260 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:27.260 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:14:27.260 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:27.260 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:27.260 [2024-11-27 15:18:55.320490] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:27.518 [2024-11-27 15:18:55.367267] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:27.518 [2024-11-27 15:18:55.367384] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:27.518 [2024-11-27 15:18:55.367450] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:27.518 [2024-11-27 15:18:55.367482] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:27.518 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:27.518 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:27.518 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:27.518 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:27.518 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:27.518 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:27.519 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:27.519 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:27.519 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:27.519 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:27.519 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:27.519 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:27.519 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:27.519 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:27.519 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:27.519 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:27.519 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:27.519 "name": "raid_bdev1", 00:14:27.519 "uuid": "a61e22e7-3726-450a-814a-3ab52053bb1c", 00:14:27.519 "strip_size_kb": 64, 00:14:27.519 "state": "online", 00:14:27.519 "raid_level": "raid5f", 00:14:27.519 "superblock": true, 00:14:27.519 "num_base_bdevs": 3, 00:14:27.519 "num_base_bdevs_discovered": 2, 00:14:27.519 "num_base_bdevs_operational": 2, 00:14:27.519 "base_bdevs_list": [ 00:14:27.519 { 00:14:27.519 "name": null, 00:14:27.519 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:27.519 "is_configured": false, 00:14:27.519 "data_offset": 0, 00:14:27.519 "data_size": 63488 00:14:27.519 }, 00:14:27.519 { 00:14:27.519 "name": "BaseBdev2", 00:14:27.519 "uuid": "975e83b2-3817-523c-8d96-fadd5982b42a", 00:14:27.519 "is_configured": true, 00:14:27.519 "data_offset": 2048, 00:14:27.519 "data_size": 63488 00:14:27.519 }, 00:14:27.519 { 00:14:27.519 "name": "BaseBdev3", 00:14:27.519 "uuid": "f9ce5ce7-36ee-5b89-b1f2-3230dcd9b374", 00:14:27.519 "is_configured": true, 00:14:27.519 "data_offset": 2048, 00:14:27.519 "data_size": 63488 00:14:27.519 } 00:14:27.519 ] 00:14:27.519 }' 00:14:27.519 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:27.519 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:27.778 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:27.778 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:27.778 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:27.778 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:27.778 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:27.778 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:27.778 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:27.778 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:27.778 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:27.778 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:27.778 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:27.778 "name": "raid_bdev1", 00:14:27.778 "uuid": "a61e22e7-3726-450a-814a-3ab52053bb1c", 00:14:27.778 "strip_size_kb": 64, 00:14:27.778 "state": "online", 00:14:27.778 "raid_level": "raid5f", 00:14:27.778 "superblock": true, 00:14:27.778 "num_base_bdevs": 3, 00:14:27.778 "num_base_bdevs_discovered": 2, 00:14:27.778 "num_base_bdevs_operational": 2, 00:14:27.778 "base_bdevs_list": [ 00:14:27.778 { 00:14:27.778 "name": null, 00:14:27.778 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:27.778 "is_configured": false, 00:14:27.778 "data_offset": 0, 00:14:27.778 "data_size": 63488 00:14:27.778 }, 00:14:27.778 { 00:14:27.778 "name": "BaseBdev2", 00:14:27.778 "uuid": "975e83b2-3817-523c-8d96-fadd5982b42a", 00:14:27.778 "is_configured": true, 00:14:27.778 "data_offset": 2048, 00:14:27.778 "data_size": 63488 00:14:27.778 }, 00:14:27.778 { 00:14:27.778 "name": "BaseBdev3", 00:14:27.778 "uuid": "f9ce5ce7-36ee-5b89-b1f2-3230dcd9b374", 00:14:27.778 "is_configured": true, 00:14:27.778 "data_offset": 2048, 00:14:27.778 "data_size": 63488 00:14:27.778 } 00:14:27.778 ] 00:14:27.778 }' 00:14:27.778 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:27.778 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:27.778 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:28.038 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:28.038 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:14:28.038 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:28.038 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:28.038 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:28.038 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:14:28.038 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:28.038 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:28.038 [2024-11-27 15:18:55.928130] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:14:28.038 [2024-11-27 15:18:55.928191] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:28.038 [2024-11-27 15:18:55.928211] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:14:28.038 [2024-11-27 15:18:55.928222] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:28.038 [2024-11-27 15:18:55.928605] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:28.038 [2024-11-27 15:18:55.928625] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:28.038 [2024-11-27 15:18:55.928687] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:14:28.038 [2024-11-27 15:18:55.928702] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:14:28.038 [2024-11-27 15:18:55.928710] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:14:28.038 [2024-11-27 15:18:55.928721] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:14:28.038 BaseBdev1 00:14:28.038 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:28.038 15:18:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:14:28.975 15:18:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:28.975 15:18:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:28.975 15:18:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:28.975 15:18:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:28.975 15:18:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:28.975 15:18:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:28.975 15:18:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:28.975 15:18:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:28.975 15:18:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:28.975 15:18:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:28.975 15:18:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:28.975 15:18:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:28.975 15:18:56 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:28.975 15:18:56 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:28.975 15:18:56 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:28.975 15:18:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:28.975 "name": "raid_bdev1", 00:14:28.975 "uuid": "a61e22e7-3726-450a-814a-3ab52053bb1c", 00:14:28.975 "strip_size_kb": 64, 00:14:28.975 "state": "online", 00:14:28.975 "raid_level": "raid5f", 00:14:28.975 "superblock": true, 00:14:28.975 "num_base_bdevs": 3, 00:14:28.975 "num_base_bdevs_discovered": 2, 00:14:28.975 "num_base_bdevs_operational": 2, 00:14:28.975 "base_bdevs_list": [ 00:14:28.975 { 00:14:28.975 "name": null, 00:14:28.975 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:28.975 "is_configured": false, 00:14:28.975 "data_offset": 0, 00:14:28.975 "data_size": 63488 00:14:28.975 }, 00:14:28.975 { 00:14:28.975 "name": "BaseBdev2", 00:14:28.975 "uuid": "975e83b2-3817-523c-8d96-fadd5982b42a", 00:14:28.975 "is_configured": true, 00:14:28.975 "data_offset": 2048, 00:14:28.975 "data_size": 63488 00:14:28.975 }, 00:14:28.975 { 00:14:28.975 "name": "BaseBdev3", 00:14:28.975 "uuid": "f9ce5ce7-36ee-5b89-b1f2-3230dcd9b374", 00:14:28.975 "is_configured": true, 00:14:28.975 "data_offset": 2048, 00:14:28.975 "data_size": 63488 00:14:28.975 } 00:14:28.975 ] 00:14:28.975 }' 00:14:28.975 15:18:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:28.975 15:18:56 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:29.545 15:18:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:29.545 15:18:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:29.545 15:18:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:29.545 15:18:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:29.545 15:18:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:29.545 15:18:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:29.545 15:18:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:29.545 15:18:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:29.545 15:18:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:29.545 15:18:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:29.545 15:18:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:29.545 "name": "raid_bdev1", 00:14:29.545 "uuid": "a61e22e7-3726-450a-814a-3ab52053bb1c", 00:14:29.545 "strip_size_kb": 64, 00:14:29.545 "state": "online", 00:14:29.545 "raid_level": "raid5f", 00:14:29.545 "superblock": true, 00:14:29.545 "num_base_bdevs": 3, 00:14:29.545 "num_base_bdevs_discovered": 2, 00:14:29.545 "num_base_bdevs_operational": 2, 00:14:29.545 "base_bdevs_list": [ 00:14:29.545 { 00:14:29.545 "name": null, 00:14:29.545 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:29.545 "is_configured": false, 00:14:29.545 "data_offset": 0, 00:14:29.545 "data_size": 63488 00:14:29.545 }, 00:14:29.545 { 00:14:29.545 "name": "BaseBdev2", 00:14:29.545 "uuid": "975e83b2-3817-523c-8d96-fadd5982b42a", 00:14:29.545 "is_configured": true, 00:14:29.545 "data_offset": 2048, 00:14:29.545 "data_size": 63488 00:14:29.545 }, 00:14:29.545 { 00:14:29.545 "name": "BaseBdev3", 00:14:29.545 "uuid": "f9ce5ce7-36ee-5b89-b1f2-3230dcd9b374", 00:14:29.545 "is_configured": true, 00:14:29.545 "data_offset": 2048, 00:14:29.545 "data_size": 63488 00:14:29.545 } 00:14:29.545 ] 00:14:29.545 }' 00:14:29.545 15:18:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:29.545 15:18:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:29.545 15:18:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:29.545 15:18:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:29.545 15:18:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:29.545 15:18:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@652 -- # local es=0 00:14:29.545 15:18:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:29.545 15:18:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:14:29.545 15:18:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:29.545 15:18:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:14:29.545 15:18:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:29.545 15:18:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:29.545 15:18:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:29.545 15:18:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:29.545 [2024-11-27 15:18:57.601509] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:29.545 [2024-11-27 15:18:57.601760] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:14:29.545 [2024-11-27 15:18:57.601819] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:14:29.545 request: 00:14:29.545 { 00:14:29.545 "base_bdev": "BaseBdev1", 00:14:29.545 "raid_bdev": "raid_bdev1", 00:14:29.545 "method": "bdev_raid_add_base_bdev", 00:14:29.545 "req_id": 1 00:14:29.545 } 00:14:29.545 Got JSON-RPC error response 00:14:29.545 response: 00:14:29.545 { 00:14:29.545 "code": -22, 00:14:29.545 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:14:29.545 } 00:14:29.545 15:18:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:14:29.545 15:18:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@655 -- # es=1 00:14:29.545 15:18:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:14:29.545 15:18:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:14:29.545 15:18:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:14:29.545 15:18:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:14:30.927 15:18:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:30.927 15:18:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:30.927 15:18:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:30.927 15:18:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:30.927 15:18:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:30.927 15:18:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:30.927 15:18:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:30.927 15:18:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:30.927 15:18:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:30.927 15:18:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:30.927 15:18:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:30.927 15:18:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:30.927 15:18:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:30.927 15:18:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:30.927 15:18:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:30.927 15:18:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:30.927 "name": "raid_bdev1", 00:14:30.927 "uuid": "a61e22e7-3726-450a-814a-3ab52053bb1c", 00:14:30.927 "strip_size_kb": 64, 00:14:30.927 "state": "online", 00:14:30.927 "raid_level": "raid5f", 00:14:30.927 "superblock": true, 00:14:30.927 "num_base_bdevs": 3, 00:14:30.927 "num_base_bdevs_discovered": 2, 00:14:30.927 "num_base_bdevs_operational": 2, 00:14:30.927 "base_bdevs_list": [ 00:14:30.927 { 00:14:30.927 "name": null, 00:14:30.927 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:30.927 "is_configured": false, 00:14:30.927 "data_offset": 0, 00:14:30.927 "data_size": 63488 00:14:30.927 }, 00:14:30.927 { 00:14:30.927 "name": "BaseBdev2", 00:14:30.927 "uuid": "975e83b2-3817-523c-8d96-fadd5982b42a", 00:14:30.927 "is_configured": true, 00:14:30.927 "data_offset": 2048, 00:14:30.927 "data_size": 63488 00:14:30.927 }, 00:14:30.927 { 00:14:30.927 "name": "BaseBdev3", 00:14:30.927 "uuid": "f9ce5ce7-36ee-5b89-b1f2-3230dcd9b374", 00:14:30.927 "is_configured": true, 00:14:30.927 "data_offset": 2048, 00:14:30.927 "data_size": 63488 00:14:30.927 } 00:14:30.927 ] 00:14:30.927 }' 00:14:30.927 15:18:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:30.927 15:18:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:31.201 15:18:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:31.201 15:18:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:31.201 15:18:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:31.201 15:18:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:31.201 15:18:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:31.201 15:18:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:31.201 15:18:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:31.201 15:18:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:31.201 15:18:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:31.202 15:18:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:31.202 15:18:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:31.202 "name": "raid_bdev1", 00:14:31.202 "uuid": "a61e22e7-3726-450a-814a-3ab52053bb1c", 00:14:31.202 "strip_size_kb": 64, 00:14:31.202 "state": "online", 00:14:31.202 "raid_level": "raid5f", 00:14:31.202 "superblock": true, 00:14:31.202 "num_base_bdevs": 3, 00:14:31.202 "num_base_bdevs_discovered": 2, 00:14:31.202 "num_base_bdevs_operational": 2, 00:14:31.202 "base_bdevs_list": [ 00:14:31.202 { 00:14:31.202 "name": null, 00:14:31.202 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:31.202 "is_configured": false, 00:14:31.202 "data_offset": 0, 00:14:31.202 "data_size": 63488 00:14:31.202 }, 00:14:31.202 { 00:14:31.202 "name": "BaseBdev2", 00:14:31.202 "uuid": "975e83b2-3817-523c-8d96-fadd5982b42a", 00:14:31.202 "is_configured": true, 00:14:31.202 "data_offset": 2048, 00:14:31.202 "data_size": 63488 00:14:31.202 }, 00:14:31.202 { 00:14:31.202 "name": "BaseBdev3", 00:14:31.202 "uuid": "f9ce5ce7-36ee-5b89-b1f2-3230dcd9b374", 00:14:31.202 "is_configured": true, 00:14:31.202 "data_offset": 2048, 00:14:31.202 "data_size": 63488 00:14:31.202 } 00:14:31.202 ] 00:14:31.202 }' 00:14:31.203 15:18:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:31.203 15:18:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:31.203 15:18:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:31.203 15:18:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:31.203 15:18:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 92699 00:14:31.203 15:18:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@954 -- # '[' -z 92699 ']' 00:14:31.203 15:18:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@958 -- # kill -0 92699 00:14:31.203 15:18:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@959 -- # uname 00:14:31.203 15:18:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:31.203 15:18:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 92699 00:14:31.203 15:18:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:14:31.203 15:18:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:14:31.204 killing process with pid 92699 00:14:31.204 Received shutdown signal, test time was about 60.000000 seconds 00:14:31.204 00:14:31.204 Latency(us) 00:14:31.204 [2024-11-27T15:18:59.311Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:14:31.204 [2024-11-27T15:18:59.311Z] =================================================================================================================== 00:14:31.204 [2024-11-27T15:18:59.311Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:14:31.204 15:18:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 92699' 00:14:31.204 15:18:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@973 -- # kill 92699 00:14:31.204 [2024-11-27 15:18:59.195865] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:31.204 15:18:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@978 -- # wait 92699 00:14:31.204 [2024-11-27 15:18:59.196039] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:31.204 [2024-11-27 15:18:59.196129] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:31.205 [2024-11-27 15:18:59.196141] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state offline 00:14:31.205 [2024-11-27 15:18:59.275498] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:31.779 15:18:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:14:31.779 00:14:31.779 real 0m21.873s 00:14:31.779 user 0m28.582s 00:14:31.779 sys 0m2.716s 00:14:31.779 15:18:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:31.779 ************************************ 00:14:31.779 END TEST raid5f_rebuild_test_sb 00:14:31.779 ************************************ 00:14:31.779 15:18:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:31.779 15:18:59 bdev_raid -- bdev/bdev_raid.sh@985 -- # for n in {3..4} 00:14:31.779 15:18:59 bdev_raid -- bdev/bdev_raid.sh@986 -- # run_test raid5f_state_function_test raid_state_function_test raid5f 4 false 00:14:31.779 15:18:59 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:14:31.779 15:18:59 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:31.779 15:18:59 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:31.779 ************************************ 00:14:31.779 START TEST raid5f_state_function_test 00:14:31.779 ************************************ 00:14:31.779 15:18:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid5f 4 false 00:14:31.779 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:14:31.779 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:14:31.779 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:14:31.779 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:14:31.779 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:14:31.779 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:31.779 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:14:31.779 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:31.779 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:31.779 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:14:31.779 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:31.779 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:31.779 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:14:31.779 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:31.779 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:31.779 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:14:31.779 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:31.779 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:31.779 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:14:31.779 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:14:31.779 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:14:31.779 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:14:31.779 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:14:31.779 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:14:31.779 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:14:31.779 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:14:31.780 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:14:31.780 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:14:31.780 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:14:31.780 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=93430 00:14:31.780 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:14:31.780 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 93430' 00:14:31.780 Process raid pid: 93430 00:14:31.780 15:18:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 93430 00:14:31.780 15:18:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 93430 ']' 00:14:31.780 15:18:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:31.780 15:18:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:31.780 15:18:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:31.780 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:31.780 15:18:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:31.780 15:18:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.780 [2024-11-27 15:18:59.761329] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:14:31.780 [2024-11-27 15:18:59.761557] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:32.039 [2024-11-27 15:18:59.931183] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:32.040 [2024-11-27 15:18:59.970715] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:32.040 [2024-11-27 15:19:00.047782] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:32.040 [2024-11-27 15:19:00.047973] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:32.609 15:19:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:32.609 15:19:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:14:32.609 15:19:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:32.609 15:19:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:32.609 15:19:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:32.609 [2024-11-27 15:19:00.614398] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:32.609 [2024-11-27 15:19:00.614545] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:32.609 [2024-11-27 15:19:00.614579] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:32.609 [2024-11-27 15:19:00.614608] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:32.609 [2024-11-27 15:19:00.614629] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:32.609 [2024-11-27 15:19:00.614667] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:32.609 [2024-11-27 15:19:00.614688] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:32.609 [2024-11-27 15:19:00.614747] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:32.609 15:19:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:32.609 15:19:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:32.609 15:19:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:32.609 15:19:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:32.609 15:19:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:32.609 15:19:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:32.609 15:19:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:32.609 15:19:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:32.609 15:19:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:32.609 15:19:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:32.610 15:19:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:32.610 15:19:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:32.610 15:19:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:32.610 15:19:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:32.610 15:19:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:32.610 15:19:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:32.610 15:19:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:32.610 "name": "Existed_Raid", 00:14:32.610 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:32.610 "strip_size_kb": 64, 00:14:32.610 "state": "configuring", 00:14:32.610 "raid_level": "raid5f", 00:14:32.610 "superblock": false, 00:14:32.610 "num_base_bdevs": 4, 00:14:32.610 "num_base_bdevs_discovered": 0, 00:14:32.610 "num_base_bdevs_operational": 4, 00:14:32.610 "base_bdevs_list": [ 00:14:32.610 { 00:14:32.610 "name": "BaseBdev1", 00:14:32.610 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:32.610 "is_configured": false, 00:14:32.610 "data_offset": 0, 00:14:32.610 "data_size": 0 00:14:32.610 }, 00:14:32.610 { 00:14:32.610 "name": "BaseBdev2", 00:14:32.610 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:32.610 "is_configured": false, 00:14:32.610 "data_offset": 0, 00:14:32.610 "data_size": 0 00:14:32.610 }, 00:14:32.610 { 00:14:32.610 "name": "BaseBdev3", 00:14:32.610 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:32.610 "is_configured": false, 00:14:32.610 "data_offset": 0, 00:14:32.610 "data_size": 0 00:14:32.610 }, 00:14:32.610 { 00:14:32.610 "name": "BaseBdev4", 00:14:32.610 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:32.610 "is_configured": false, 00:14:32.610 "data_offset": 0, 00:14:32.610 "data_size": 0 00:14:32.610 } 00:14:32.610 ] 00:14:32.610 }' 00:14:32.610 15:19:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:32.610 15:19:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.179 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:33.179 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.179 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.179 [2024-11-27 15:19:01.065527] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:33.179 [2024-11-27 15:19:01.065623] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:14:33.179 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.179 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:33.179 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.179 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.179 [2024-11-27 15:19:01.077525] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:33.179 [2024-11-27 15:19:01.077617] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:33.179 [2024-11-27 15:19:01.077647] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:33.179 [2024-11-27 15:19:01.077674] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:33.179 [2024-11-27 15:19:01.077695] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:33.179 [2024-11-27 15:19:01.077720] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:33.179 [2024-11-27 15:19:01.077741] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:33.179 [2024-11-27 15:19:01.077766] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:33.179 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.179 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:33.179 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.179 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.179 [2024-11-27 15:19:01.104498] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:33.179 BaseBdev1 00:14:33.179 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.179 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:14:33.179 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:14:33.179 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:33.179 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:33.179 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:33.179 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:33.179 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:33.179 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.179 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.179 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.179 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:33.179 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.179 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.179 [ 00:14:33.179 { 00:14:33.179 "name": "BaseBdev1", 00:14:33.179 "aliases": [ 00:14:33.179 "9d3078a4-21bc-40ae-a56b-73a18ff58b6e" 00:14:33.179 ], 00:14:33.179 "product_name": "Malloc disk", 00:14:33.179 "block_size": 512, 00:14:33.179 "num_blocks": 65536, 00:14:33.179 "uuid": "9d3078a4-21bc-40ae-a56b-73a18ff58b6e", 00:14:33.179 "assigned_rate_limits": { 00:14:33.179 "rw_ios_per_sec": 0, 00:14:33.179 "rw_mbytes_per_sec": 0, 00:14:33.179 "r_mbytes_per_sec": 0, 00:14:33.179 "w_mbytes_per_sec": 0 00:14:33.179 }, 00:14:33.179 "claimed": true, 00:14:33.179 "claim_type": "exclusive_write", 00:14:33.179 "zoned": false, 00:14:33.179 "supported_io_types": { 00:14:33.179 "read": true, 00:14:33.179 "write": true, 00:14:33.179 "unmap": true, 00:14:33.179 "flush": true, 00:14:33.179 "reset": true, 00:14:33.179 "nvme_admin": false, 00:14:33.179 "nvme_io": false, 00:14:33.179 "nvme_io_md": false, 00:14:33.179 "write_zeroes": true, 00:14:33.179 "zcopy": true, 00:14:33.179 "get_zone_info": false, 00:14:33.179 "zone_management": false, 00:14:33.179 "zone_append": false, 00:14:33.179 "compare": false, 00:14:33.179 "compare_and_write": false, 00:14:33.179 "abort": true, 00:14:33.179 "seek_hole": false, 00:14:33.179 "seek_data": false, 00:14:33.179 "copy": true, 00:14:33.179 "nvme_iov_md": false 00:14:33.179 }, 00:14:33.179 "memory_domains": [ 00:14:33.179 { 00:14:33.179 "dma_device_id": "system", 00:14:33.179 "dma_device_type": 1 00:14:33.179 }, 00:14:33.179 { 00:14:33.179 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:33.179 "dma_device_type": 2 00:14:33.179 } 00:14:33.179 ], 00:14:33.179 "driver_specific": {} 00:14:33.179 } 00:14:33.179 ] 00:14:33.180 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.180 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:33.180 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:33.180 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:33.180 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:33.180 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:33.180 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:33.180 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:33.180 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:33.180 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:33.180 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:33.180 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:33.180 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:33.180 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:33.180 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.180 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.180 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.180 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:33.180 "name": "Existed_Raid", 00:14:33.180 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:33.180 "strip_size_kb": 64, 00:14:33.180 "state": "configuring", 00:14:33.180 "raid_level": "raid5f", 00:14:33.180 "superblock": false, 00:14:33.180 "num_base_bdevs": 4, 00:14:33.180 "num_base_bdevs_discovered": 1, 00:14:33.180 "num_base_bdevs_operational": 4, 00:14:33.180 "base_bdevs_list": [ 00:14:33.180 { 00:14:33.180 "name": "BaseBdev1", 00:14:33.180 "uuid": "9d3078a4-21bc-40ae-a56b-73a18ff58b6e", 00:14:33.180 "is_configured": true, 00:14:33.180 "data_offset": 0, 00:14:33.180 "data_size": 65536 00:14:33.180 }, 00:14:33.180 { 00:14:33.180 "name": "BaseBdev2", 00:14:33.180 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:33.180 "is_configured": false, 00:14:33.180 "data_offset": 0, 00:14:33.180 "data_size": 0 00:14:33.180 }, 00:14:33.180 { 00:14:33.180 "name": "BaseBdev3", 00:14:33.180 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:33.180 "is_configured": false, 00:14:33.180 "data_offset": 0, 00:14:33.180 "data_size": 0 00:14:33.180 }, 00:14:33.180 { 00:14:33.180 "name": "BaseBdev4", 00:14:33.180 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:33.180 "is_configured": false, 00:14:33.180 "data_offset": 0, 00:14:33.180 "data_size": 0 00:14:33.180 } 00:14:33.180 ] 00:14:33.180 }' 00:14:33.180 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:33.180 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.750 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:33.750 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.750 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.750 [2024-11-27 15:19:01.559695] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:33.750 [2024-11-27 15:19:01.559793] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:14:33.750 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.750 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:33.750 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.750 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.750 [2024-11-27 15:19:01.571723] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:33.750 [2024-11-27 15:19:01.573838] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:33.750 [2024-11-27 15:19:01.573888] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:33.750 [2024-11-27 15:19:01.573911] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:33.750 [2024-11-27 15:19:01.573922] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:33.750 [2024-11-27 15:19:01.573930] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:33.750 [2024-11-27 15:19:01.573941] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:33.750 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.750 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:14:33.750 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:33.750 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:33.750 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:33.750 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:33.750 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:33.750 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:33.750 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:33.750 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:33.750 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:33.750 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:33.750 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:33.750 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:33.750 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:33.750 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.750 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.750 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.750 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:33.750 "name": "Existed_Raid", 00:14:33.750 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:33.750 "strip_size_kb": 64, 00:14:33.750 "state": "configuring", 00:14:33.750 "raid_level": "raid5f", 00:14:33.750 "superblock": false, 00:14:33.750 "num_base_bdevs": 4, 00:14:33.750 "num_base_bdevs_discovered": 1, 00:14:33.750 "num_base_bdevs_operational": 4, 00:14:33.750 "base_bdevs_list": [ 00:14:33.750 { 00:14:33.750 "name": "BaseBdev1", 00:14:33.750 "uuid": "9d3078a4-21bc-40ae-a56b-73a18ff58b6e", 00:14:33.750 "is_configured": true, 00:14:33.750 "data_offset": 0, 00:14:33.750 "data_size": 65536 00:14:33.750 }, 00:14:33.750 { 00:14:33.750 "name": "BaseBdev2", 00:14:33.750 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:33.750 "is_configured": false, 00:14:33.750 "data_offset": 0, 00:14:33.750 "data_size": 0 00:14:33.750 }, 00:14:33.750 { 00:14:33.750 "name": "BaseBdev3", 00:14:33.750 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:33.750 "is_configured": false, 00:14:33.750 "data_offset": 0, 00:14:33.750 "data_size": 0 00:14:33.750 }, 00:14:33.750 { 00:14:33.750 "name": "BaseBdev4", 00:14:33.750 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:33.750 "is_configured": false, 00:14:33.750 "data_offset": 0, 00:14:33.750 "data_size": 0 00:14:33.750 } 00:14:33.750 ] 00:14:33.750 }' 00:14:33.750 15:19:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:33.750 15:19:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:34.010 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:34.010 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:34.010 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:34.010 [2024-11-27 15:19:02.067872] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:34.010 BaseBdev2 00:14:34.010 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:34.010 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:14:34.010 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:14:34.010 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:34.010 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:34.010 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:34.010 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:34.010 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:34.010 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:34.010 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:34.010 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:34.010 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:34.010 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:34.010 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:34.010 [ 00:14:34.010 { 00:14:34.010 "name": "BaseBdev2", 00:14:34.010 "aliases": [ 00:14:34.010 "036b8da2-894e-4991-86e5-b6e876d0088c" 00:14:34.010 ], 00:14:34.010 "product_name": "Malloc disk", 00:14:34.010 "block_size": 512, 00:14:34.010 "num_blocks": 65536, 00:14:34.010 "uuid": "036b8da2-894e-4991-86e5-b6e876d0088c", 00:14:34.010 "assigned_rate_limits": { 00:14:34.010 "rw_ios_per_sec": 0, 00:14:34.010 "rw_mbytes_per_sec": 0, 00:14:34.010 "r_mbytes_per_sec": 0, 00:14:34.010 "w_mbytes_per_sec": 0 00:14:34.010 }, 00:14:34.010 "claimed": true, 00:14:34.010 "claim_type": "exclusive_write", 00:14:34.010 "zoned": false, 00:14:34.010 "supported_io_types": { 00:14:34.010 "read": true, 00:14:34.010 "write": true, 00:14:34.010 "unmap": true, 00:14:34.010 "flush": true, 00:14:34.010 "reset": true, 00:14:34.010 "nvme_admin": false, 00:14:34.010 "nvme_io": false, 00:14:34.010 "nvme_io_md": false, 00:14:34.010 "write_zeroes": true, 00:14:34.010 "zcopy": true, 00:14:34.010 "get_zone_info": false, 00:14:34.011 "zone_management": false, 00:14:34.011 "zone_append": false, 00:14:34.011 "compare": false, 00:14:34.011 "compare_and_write": false, 00:14:34.011 "abort": true, 00:14:34.011 "seek_hole": false, 00:14:34.011 "seek_data": false, 00:14:34.011 "copy": true, 00:14:34.011 "nvme_iov_md": false 00:14:34.011 }, 00:14:34.011 "memory_domains": [ 00:14:34.011 { 00:14:34.011 "dma_device_id": "system", 00:14:34.011 "dma_device_type": 1 00:14:34.011 }, 00:14:34.011 { 00:14:34.011 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:34.011 "dma_device_type": 2 00:14:34.011 } 00:14:34.011 ], 00:14:34.011 "driver_specific": {} 00:14:34.011 } 00:14:34.011 ] 00:14:34.011 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:34.011 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:34.011 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:34.011 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:34.011 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:34.011 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:34.011 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:34.011 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:34.011 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:34.011 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:34.011 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:34.011 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:34.011 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:34.011 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:34.011 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:34.011 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:34.011 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:34.011 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:34.271 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:34.271 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:34.271 "name": "Existed_Raid", 00:14:34.271 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:34.271 "strip_size_kb": 64, 00:14:34.271 "state": "configuring", 00:14:34.271 "raid_level": "raid5f", 00:14:34.271 "superblock": false, 00:14:34.271 "num_base_bdevs": 4, 00:14:34.271 "num_base_bdevs_discovered": 2, 00:14:34.271 "num_base_bdevs_operational": 4, 00:14:34.271 "base_bdevs_list": [ 00:14:34.271 { 00:14:34.271 "name": "BaseBdev1", 00:14:34.271 "uuid": "9d3078a4-21bc-40ae-a56b-73a18ff58b6e", 00:14:34.271 "is_configured": true, 00:14:34.271 "data_offset": 0, 00:14:34.271 "data_size": 65536 00:14:34.271 }, 00:14:34.271 { 00:14:34.271 "name": "BaseBdev2", 00:14:34.271 "uuid": "036b8da2-894e-4991-86e5-b6e876d0088c", 00:14:34.271 "is_configured": true, 00:14:34.271 "data_offset": 0, 00:14:34.271 "data_size": 65536 00:14:34.271 }, 00:14:34.271 { 00:14:34.271 "name": "BaseBdev3", 00:14:34.271 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:34.271 "is_configured": false, 00:14:34.271 "data_offset": 0, 00:14:34.271 "data_size": 0 00:14:34.271 }, 00:14:34.271 { 00:14:34.271 "name": "BaseBdev4", 00:14:34.271 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:34.271 "is_configured": false, 00:14:34.271 "data_offset": 0, 00:14:34.271 "data_size": 0 00:14:34.271 } 00:14:34.271 ] 00:14:34.271 }' 00:14:34.271 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:34.271 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:34.532 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:34.532 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:34.532 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:34.532 [2024-11-27 15:19:02.604418] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:34.532 BaseBdev3 00:14:34.532 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:34.532 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:14:34.532 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:14:34.532 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:34.532 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:34.532 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:34.532 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:34.532 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:34.532 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:34.532 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:34.532 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:34.532 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:34.532 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:34.532 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:34.532 [ 00:14:34.532 { 00:14:34.532 "name": "BaseBdev3", 00:14:34.532 "aliases": [ 00:14:34.532 "024a56eb-c5ac-4a00-9e25-b742b2930367" 00:14:34.532 ], 00:14:34.532 "product_name": "Malloc disk", 00:14:34.532 "block_size": 512, 00:14:34.532 "num_blocks": 65536, 00:14:34.532 "uuid": "024a56eb-c5ac-4a00-9e25-b742b2930367", 00:14:34.532 "assigned_rate_limits": { 00:14:34.532 "rw_ios_per_sec": 0, 00:14:34.532 "rw_mbytes_per_sec": 0, 00:14:34.532 "r_mbytes_per_sec": 0, 00:14:34.532 "w_mbytes_per_sec": 0 00:14:34.532 }, 00:14:34.532 "claimed": true, 00:14:34.532 "claim_type": "exclusive_write", 00:14:34.532 "zoned": false, 00:14:34.532 "supported_io_types": { 00:14:34.532 "read": true, 00:14:34.532 "write": true, 00:14:34.794 "unmap": true, 00:14:34.794 "flush": true, 00:14:34.794 "reset": true, 00:14:34.794 "nvme_admin": false, 00:14:34.794 "nvme_io": false, 00:14:34.794 "nvme_io_md": false, 00:14:34.794 "write_zeroes": true, 00:14:34.794 "zcopy": true, 00:14:34.794 "get_zone_info": false, 00:14:34.794 "zone_management": false, 00:14:34.794 "zone_append": false, 00:14:34.794 "compare": false, 00:14:34.794 "compare_and_write": false, 00:14:34.794 "abort": true, 00:14:34.794 "seek_hole": false, 00:14:34.794 "seek_data": false, 00:14:34.794 "copy": true, 00:14:34.794 "nvme_iov_md": false 00:14:34.794 }, 00:14:34.794 "memory_domains": [ 00:14:34.794 { 00:14:34.794 "dma_device_id": "system", 00:14:34.794 "dma_device_type": 1 00:14:34.794 }, 00:14:34.794 { 00:14:34.794 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:34.794 "dma_device_type": 2 00:14:34.794 } 00:14:34.794 ], 00:14:34.794 "driver_specific": {} 00:14:34.794 } 00:14:34.794 ] 00:14:34.794 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:34.794 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:34.794 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:34.794 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:34.794 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:34.794 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:34.794 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:34.794 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:34.794 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:34.794 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:34.794 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:34.794 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:34.794 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:34.794 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:34.794 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:34.794 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:34.794 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:34.794 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:34.794 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:34.794 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:34.794 "name": "Existed_Raid", 00:14:34.794 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:34.794 "strip_size_kb": 64, 00:14:34.794 "state": "configuring", 00:14:34.794 "raid_level": "raid5f", 00:14:34.794 "superblock": false, 00:14:34.794 "num_base_bdevs": 4, 00:14:34.794 "num_base_bdevs_discovered": 3, 00:14:34.794 "num_base_bdevs_operational": 4, 00:14:34.794 "base_bdevs_list": [ 00:14:34.794 { 00:14:34.794 "name": "BaseBdev1", 00:14:34.794 "uuid": "9d3078a4-21bc-40ae-a56b-73a18ff58b6e", 00:14:34.794 "is_configured": true, 00:14:34.794 "data_offset": 0, 00:14:34.794 "data_size": 65536 00:14:34.794 }, 00:14:34.794 { 00:14:34.794 "name": "BaseBdev2", 00:14:34.794 "uuid": "036b8da2-894e-4991-86e5-b6e876d0088c", 00:14:34.794 "is_configured": true, 00:14:34.794 "data_offset": 0, 00:14:34.794 "data_size": 65536 00:14:34.794 }, 00:14:34.794 { 00:14:34.794 "name": "BaseBdev3", 00:14:34.794 "uuid": "024a56eb-c5ac-4a00-9e25-b742b2930367", 00:14:34.794 "is_configured": true, 00:14:34.794 "data_offset": 0, 00:14:34.794 "data_size": 65536 00:14:34.794 }, 00:14:34.794 { 00:14:34.794 "name": "BaseBdev4", 00:14:34.794 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:34.794 "is_configured": false, 00:14:34.794 "data_offset": 0, 00:14:34.794 "data_size": 0 00:14:34.794 } 00:14:34.794 ] 00:14:34.794 }' 00:14:34.794 15:19:02 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:34.794 15:19:02 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.063 [2024-11-27 15:19:03.104606] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:35.063 [2024-11-27 15:19:03.104752] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:14:35.063 [2024-11-27 15:19:03.104781] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:14:35.063 [2024-11-27 15:19:03.105166] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:14:35.063 [2024-11-27 15:19:03.105700] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:14:35.063 [2024-11-27 15:19:03.105726] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:14:35.063 [2024-11-27 15:19:03.105994] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:35.063 BaseBdev4 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.063 [ 00:14:35.063 { 00:14:35.063 "name": "BaseBdev4", 00:14:35.063 "aliases": [ 00:14:35.063 "c9fe2fdf-7bb2-48e6-8af0-78ada329da1e" 00:14:35.063 ], 00:14:35.063 "product_name": "Malloc disk", 00:14:35.063 "block_size": 512, 00:14:35.063 "num_blocks": 65536, 00:14:35.063 "uuid": "c9fe2fdf-7bb2-48e6-8af0-78ada329da1e", 00:14:35.063 "assigned_rate_limits": { 00:14:35.063 "rw_ios_per_sec": 0, 00:14:35.063 "rw_mbytes_per_sec": 0, 00:14:35.063 "r_mbytes_per_sec": 0, 00:14:35.063 "w_mbytes_per_sec": 0 00:14:35.063 }, 00:14:35.063 "claimed": true, 00:14:35.063 "claim_type": "exclusive_write", 00:14:35.063 "zoned": false, 00:14:35.063 "supported_io_types": { 00:14:35.063 "read": true, 00:14:35.063 "write": true, 00:14:35.063 "unmap": true, 00:14:35.063 "flush": true, 00:14:35.063 "reset": true, 00:14:35.063 "nvme_admin": false, 00:14:35.063 "nvme_io": false, 00:14:35.063 "nvme_io_md": false, 00:14:35.063 "write_zeroes": true, 00:14:35.063 "zcopy": true, 00:14:35.063 "get_zone_info": false, 00:14:35.063 "zone_management": false, 00:14:35.063 "zone_append": false, 00:14:35.063 "compare": false, 00:14:35.063 "compare_and_write": false, 00:14:35.063 "abort": true, 00:14:35.063 "seek_hole": false, 00:14:35.063 "seek_data": false, 00:14:35.063 "copy": true, 00:14:35.063 "nvme_iov_md": false 00:14:35.063 }, 00:14:35.063 "memory_domains": [ 00:14:35.063 { 00:14:35.063 "dma_device_id": "system", 00:14:35.063 "dma_device_type": 1 00:14:35.063 }, 00:14:35.063 { 00:14:35.063 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:35.063 "dma_device_type": 2 00:14:35.063 } 00:14:35.063 ], 00:14:35.063 "driver_specific": {} 00:14:35.063 } 00:14:35.063 ] 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.063 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:35.344 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:35.344 "name": "Existed_Raid", 00:14:35.344 "uuid": "5dc965de-37c5-4d73-9c73-17387ecb8bc4", 00:14:35.344 "strip_size_kb": 64, 00:14:35.344 "state": "online", 00:14:35.344 "raid_level": "raid5f", 00:14:35.344 "superblock": false, 00:14:35.344 "num_base_bdevs": 4, 00:14:35.344 "num_base_bdevs_discovered": 4, 00:14:35.344 "num_base_bdevs_operational": 4, 00:14:35.344 "base_bdevs_list": [ 00:14:35.344 { 00:14:35.344 "name": "BaseBdev1", 00:14:35.344 "uuid": "9d3078a4-21bc-40ae-a56b-73a18ff58b6e", 00:14:35.344 "is_configured": true, 00:14:35.344 "data_offset": 0, 00:14:35.344 "data_size": 65536 00:14:35.344 }, 00:14:35.344 { 00:14:35.344 "name": "BaseBdev2", 00:14:35.344 "uuid": "036b8da2-894e-4991-86e5-b6e876d0088c", 00:14:35.344 "is_configured": true, 00:14:35.344 "data_offset": 0, 00:14:35.344 "data_size": 65536 00:14:35.344 }, 00:14:35.344 { 00:14:35.344 "name": "BaseBdev3", 00:14:35.344 "uuid": "024a56eb-c5ac-4a00-9e25-b742b2930367", 00:14:35.344 "is_configured": true, 00:14:35.344 "data_offset": 0, 00:14:35.344 "data_size": 65536 00:14:35.344 }, 00:14:35.344 { 00:14:35.344 "name": "BaseBdev4", 00:14:35.344 "uuid": "c9fe2fdf-7bb2-48e6-8af0-78ada329da1e", 00:14:35.344 "is_configured": true, 00:14:35.344 "data_offset": 0, 00:14:35.344 "data_size": 65536 00:14:35.344 } 00:14:35.344 ] 00:14:35.344 }' 00:14:35.344 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:35.344 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.614 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:14:35.614 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:35.614 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:35.614 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:35.614 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:35.614 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:35.614 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:35.614 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:35.614 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.614 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:35.614 [2024-11-27 15:19:03.544332] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:35.614 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:35.614 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:35.614 "name": "Existed_Raid", 00:14:35.614 "aliases": [ 00:14:35.614 "5dc965de-37c5-4d73-9c73-17387ecb8bc4" 00:14:35.614 ], 00:14:35.614 "product_name": "Raid Volume", 00:14:35.614 "block_size": 512, 00:14:35.614 "num_blocks": 196608, 00:14:35.614 "uuid": "5dc965de-37c5-4d73-9c73-17387ecb8bc4", 00:14:35.614 "assigned_rate_limits": { 00:14:35.614 "rw_ios_per_sec": 0, 00:14:35.614 "rw_mbytes_per_sec": 0, 00:14:35.614 "r_mbytes_per_sec": 0, 00:14:35.614 "w_mbytes_per_sec": 0 00:14:35.614 }, 00:14:35.614 "claimed": false, 00:14:35.614 "zoned": false, 00:14:35.614 "supported_io_types": { 00:14:35.614 "read": true, 00:14:35.614 "write": true, 00:14:35.614 "unmap": false, 00:14:35.614 "flush": false, 00:14:35.614 "reset": true, 00:14:35.614 "nvme_admin": false, 00:14:35.614 "nvme_io": false, 00:14:35.614 "nvme_io_md": false, 00:14:35.614 "write_zeroes": true, 00:14:35.614 "zcopy": false, 00:14:35.614 "get_zone_info": false, 00:14:35.614 "zone_management": false, 00:14:35.614 "zone_append": false, 00:14:35.614 "compare": false, 00:14:35.614 "compare_and_write": false, 00:14:35.614 "abort": false, 00:14:35.614 "seek_hole": false, 00:14:35.614 "seek_data": false, 00:14:35.614 "copy": false, 00:14:35.614 "nvme_iov_md": false 00:14:35.614 }, 00:14:35.614 "driver_specific": { 00:14:35.614 "raid": { 00:14:35.614 "uuid": "5dc965de-37c5-4d73-9c73-17387ecb8bc4", 00:14:35.614 "strip_size_kb": 64, 00:14:35.614 "state": "online", 00:14:35.614 "raid_level": "raid5f", 00:14:35.614 "superblock": false, 00:14:35.614 "num_base_bdevs": 4, 00:14:35.614 "num_base_bdevs_discovered": 4, 00:14:35.614 "num_base_bdevs_operational": 4, 00:14:35.614 "base_bdevs_list": [ 00:14:35.614 { 00:14:35.614 "name": "BaseBdev1", 00:14:35.614 "uuid": "9d3078a4-21bc-40ae-a56b-73a18ff58b6e", 00:14:35.614 "is_configured": true, 00:14:35.614 "data_offset": 0, 00:14:35.614 "data_size": 65536 00:14:35.614 }, 00:14:35.614 { 00:14:35.614 "name": "BaseBdev2", 00:14:35.614 "uuid": "036b8da2-894e-4991-86e5-b6e876d0088c", 00:14:35.614 "is_configured": true, 00:14:35.614 "data_offset": 0, 00:14:35.614 "data_size": 65536 00:14:35.614 }, 00:14:35.614 { 00:14:35.614 "name": "BaseBdev3", 00:14:35.614 "uuid": "024a56eb-c5ac-4a00-9e25-b742b2930367", 00:14:35.614 "is_configured": true, 00:14:35.614 "data_offset": 0, 00:14:35.614 "data_size": 65536 00:14:35.614 }, 00:14:35.614 { 00:14:35.614 "name": "BaseBdev4", 00:14:35.614 "uuid": "c9fe2fdf-7bb2-48e6-8af0-78ada329da1e", 00:14:35.614 "is_configured": true, 00:14:35.614 "data_offset": 0, 00:14:35.614 "data_size": 65536 00:14:35.614 } 00:14:35.614 ] 00:14:35.614 } 00:14:35.614 } 00:14:35.614 }' 00:14:35.615 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:35.615 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:14:35.615 BaseBdev2 00:14:35.615 BaseBdev3 00:14:35.615 BaseBdev4' 00:14:35.615 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:35.615 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:35.615 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:35.615 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:14:35.615 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:35.615 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:35.615 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.615 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.875 [2024-11-27 15:19:03.871625] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:35.875 "name": "Existed_Raid", 00:14:35.875 "uuid": "5dc965de-37c5-4d73-9c73-17387ecb8bc4", 00:14:35.875 "strip_size_kb": 64, 00:14:35.875 "state": "online", 00:14:35.875 "raid_level": "raid5f", 00:14:35.875 "superblock": false, 00:14:35.875 "num_base_bdevs": 4, 00:14:35.875 "num_base_bdevs_discovered": 3, 00:14:35.875 "num_base_bdevs_operational": 3, 00:14:35.875 "base_bdevs_list": [ 00:14:35.875 { 00:14:35.875 "name": null, 00:14:35.875 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:35.875 "is_configured": false, 00:14:35.875 "data_offset": 0, 00:14:35.875 "data_size": 65536 00:14:35.875 }, 00:14:35.875 { 00:14:35.875 "name": "BaseBdev2", 00:14:35.875 "uuid": "036b8da2-894e-4991-86e5-b6e876d0088c", 00:14:35.875 "is_configured": true, 00:14:35.875 "data_offset": 0, 00:14:35.875 "data_size": 65536 00:14:35.875 }, 00:14:35.875 { 00:14:35.875 "name": "BaseBdev3", 00:14:35.875 "uuid": "024a56eb-c5ac-4a00-9e25-b742b2930367", 00:14:35.875 "is_configured": true, 00:14:35.875 "data_offset": 0, 00:14:35.875 "data_size": 65536 00:14:35.875 }, 00:14:35.875 { 00:14:35.875 "name": "BaseBdev4", 00:14:35.875 "uuid": "c9fe2fdf-7bb2-48e6-8af0-78ada329da1e", 00:14:35.875 "is_configured": true, 00:14:35.875 "data_offset": 0, 00:14:35.875 "data_size": 65536 00:14:35.875 } 00:14:35.875 ] 00:14:35.875 }' 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:35.875 15:19:03 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:36.445 [2024-11-27 15:19:04.395999] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:36.445 [2024-11-27 15:19:04.396174] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:36.445 [2024-11-27 15:19:04.417024] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:36.445 [2024-11-27 15:19:04.476850] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:36.445 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:36.706 [2024-11-27 15:19:04.557207] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:14:36.706 [2024-11-27 15:19:04.557332] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:36.706 BaseBdev2 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:36.706 [ 00:14:36.706 { 00:14:36.706 "name": "BaseBdev2", 00:14:36.706 "aliases": [ 00:14:36.706 "eb742695-1fb8-4278-affc-c00aeeda44e3" 00:14:36.706 ], 00:14:36.706 "product_name": "Malloc disk", 00:14:36.706 "block_size": 512, 00:14:36.706 "num_blocks": 65536, 00:14:36.706 "uuid": "eb742695-1fb8-4278-affc-c00aeeda44e3", 00:14:36.706 "assigned_rate_limits": { 00:14:36.706 "rw_ios_per_sec": 0, 00:14:36.706 "rw_mbytes_per_sec": 0, 00:14:36.706 "r_mbytes_per_sec": 0, 00:14:36.706 "w_mbytes_per_sec": 0 00:14:36.706 }, 00:14:36.706 "claimed": false, 00:14:36.706 "zoned": false, 00:14:36.706 "supported_io_types": { 00:14:36.706 "read": true, 00:14:36.706 "write": true, 00:14:36.706 "unmap": true, 00:14:36.706 "flush": true, 00:14:36.706 "reset": true, 00:14:36.706 "nvme_admin": false, 00:14:36.706 "nvme_io": false, 00:14:36.706 "nvme_io_md": false, 00:14:36.706 "write_zeroes": true, 00:14:36.706 "zcopy": true, 00:14:36.706 "get_zone_info": false, 00:14:36.706 "zone_management": false, 00:14:36.706 "zone_append": false, 00:14:36.706 "compare": false, 00:14:36.706 "compare_and_write": false, 00:14:36.706 "abort": true, 00:14:36.706 "seek_hole": false, 00:14:36.706 "seek_data": false, 00:14:36.706 "copy": true, 00:14:36.706 "nvme_iov_md": false 00:14:36.706 }, 00:14:36.706 "memory_domains": [ 00:14:36.706 { 00:14:36.706 "dma_device_id": "system", 00:14:36.706 "dma_device_type": 1 00:14:36.706 }, 00:14:36.706 { 00:14:36.706 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:36.706 "dma_device_type": 2 00:14:36.706 } 00:14:36.706 ], 00:14:36.706 "driver_specific": {} 00:14:36.706 } 00:14:36.706 ] 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:36.706 BaseBdev3 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.706 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:36.707 [ 00:14:36.707 { 00:14:36.707 "name": "BaseBdev3", 00:14:36.707 "aliases": [ 00:14:36.707 "ffd4df99-7b13-4687-bcb5-b3ec1525f886" 00:14:36.707 ], 00:14:36.707 "product_name": "Malloc disk", 00:14:36.707 "block_size": 512, 00:14:36.707 "num_blocks": 65536, 00:14:36.707 "uuid": "ffd4df99-7b13-4687-bcb5-b3ec1525f886", 00:14:36.707 "assigned_rate_limits": { 00:14:36.707 "rw_ios_per_sec": 0, 00:14:36.707 "rw_mbytes_per_sec": 0, 00:14:36.707 "r_mbytes_per_sec": 0, 00:14:36.707 "w_mbytes_per_sec": 0 00:14:36.707 }, 00:14:36.707 "claimed": false, 00:14:36.707 "zoned": false, 00:14:36.707 "supported_io_types": { 00:14:36.707 "read": true, 00:14:36.707 "write": true, 00:14:36.707 "unmap": true, 00:14:36.707 "flush": true, 00:14:36.707 "reset": true, 00:14:36.707 "nvme_admin": false, 00:14:36.707 "nvme_io": false, 00:14:36.707 "nvme_io_md": false, 00:14:36.707 "write_zeroes": true, 00:14:36.707 "zcopy": true, 00:14:36.707 "get_zone_info": false, 00:14:36.707 "zone_management": false, 00:14:36.707 "zone_append": false, 00:14:36.707 "compare": false, 00:14:36.707 "compare_and_write": false, 00:14:36.707 "abort": true, 00:14:36.707 "seek_hole": false, 00:14:36.707 "seek_data": false, 00:14:36.707 "copy": true, 00:14:36.707 "nvme_iov_md": false 00:14:36.707 }, 00:14:36.707 "memory_domains": [ 00:14:36.707 { 00:14:36.707 "dma_device_id": "system", 00:14:36.707 "dma_device_type": 1 00:14:36.707 }, 00:14:36.707 { 00:14:36.707 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:36.707 "dma_device_type": 2 00:14:36.707 } 00:14:36.707 ], 00:14:36.707 "driver_specific": {} 00:14:36.707 } 00:14:36.707 ] 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:36.707 BaseBdev4 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:36.707 [ 00:14:36.707 { 00:14:36.707 "name": "BaseBdev4", 00:14:36.707 "aliases": [ 00:14:36.707 "3c7bbd24-5ff0-4c4b-9ede-e4f9a28a1ec0" 00:14:36.707 ], 00:14:36.707 "product_name": "Malloc disk", 00:14:36.707 "block_size": 512, 00:14:36.707 "num_blocks": 65536, 00:14:36.707 "uuid": "3c7bbd24-5ff0-4c4b-9ede-e4f9a28a1ec0", 00:14:36.707 "assigned_rate_limits": { 00:14:36.707 "rw_ios_per_sec": 0, 00:14:36.707 "rw_mbytes_per_sec": 0, 00:14:36.707 "r_mbytes_per_sec": 0, 00:14:36.707 "w_mbytes_per_sec": 0 00:14:36.707 }, 00:14:36.707 "claimed": false, 00:14:36.707 "zoned": false, 00:14:36.707 "supported_io_types": { 00:14:36.707 "read": true, 00:14:36.707 "write": true, 00:14:36.707 "unmap": true, 00:14:36.707 "flush": true, 00:14:36.707 "reset": true, 00:14:36.707 "nvme_admin": false, 00:14:36.707 "nvme_io": false, 00:14:36.707 "nvme_io_md": false, 00:14:36.707 "write_zeroes": true, 00:14:36.707 "zcopy": true, 00:14:36.707 "get_zone_info": false, 00:14:36.707 "zone_management": false, 00:14:36.707 "zone_append": false, 00:14:36.707 "compare": false, 00:14:36.707 "compare_and_write": false, 00:14:36.707 "abort": true, 00:14:36.707 "seek_hole": false, 00:14:36.707 "seek_data": false, 00:14:36.707 "copy": true, 00:14:36.707 "nvme_iov_md": false 00:14:36.707 }, 00:14:36.707 "memory_domains": [ 00:14:36.707 { 00:14:36.707 "dma_device_id": "system", 00:14:36.707 "dma_device_type": 1 00:14:36.707 }, 00:14:36.707 { 00:14:36.707 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:36.707 "dma_device_type": 2 00:14:36.707 } 00:14:36.707 ], 00:14:36.707 "driver_specific": {} 00:14:36.707 } 00:14:36.707 ] 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.707 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:36.967 [2024-11-27 15:19:04.812886] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:36.967 [2024-11-27 15:19:04.813045] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:36.967 [2024-11-27 15:19:04.813096] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:36.967 [2024-11-27 15:19:04.815222] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:36.967 [2024-11-27 15:19:04.815330] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:36.967 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.967 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:36.967 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:36.967 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:36.967 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:36.967 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:36.967 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:36.967 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:36.967 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:36.967 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:36.967 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:36.967 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:36.967 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.967 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:36.967 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:36.967 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.967 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:36.967 "name": "Existed_Raid", 00:14:36.967 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:36.967 "strip_size_kb": 64, 00:14:36.967 "state": "configuring", 00:14:36.967 "raid_level": "raid5f", 00:14:36.967 "superblock": false, 00:14:36.967 "num_base_bdevs": 4, 00:14:36.967 "num_base_bdevs_discovered": 3, 00:14:36.967 "num_base_bdevs_operational": 4, 00:14:36.967 "base_bdevs_list": [ 00:14:36.967 { 00:14:36.967 "name": "BaseBdev1", 00:14:36.967 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:36.967 "is_configured": false, 00:14:36.967 "data_offset": 0, 00:14:36.967 "data_size": 0 00:14:36.967 }, 00:14:36.967 { 00:14:36.967 "name": "BaseBdev2", 00:14:36.967 "uuid": "eb742695-1fb8-4278-affc-c00aeeda44e3", 00:14:36.967 "is_configured": true, 00:14:36.967 "data_offset": 0, 00:14:36.967 "data_size": 65536 00:14:36.967 }, 00:14:36.967 { 00:14:36.967 "name": "BaseBdev3", 00:14:36.967 "uuid": "ffd4df99-7b13-4687-bcb5-b3ec1525f886", 00:14:36.967 "is_configured": true, 00:14:36.967 "data_offset": 0, 00:14:36.967 "data_size": 65536 00:14:36.967 }, 00:14:36.967 { 00:14:36.967 "name": "BaseBdev4", 00:14:36.967 "uuid": "3c7bbd24-5ff0-4c4b-9ede-e4f9a28a1ec0", 00:14:36.967 "is_configured": true, 00:14:36.967 "data_offset": 0, 00:14:36.967 "data_size": 65536 00:14:36.967 } 00:14:36.967 ] 00:14:36.967 }' 00:14:36.967 15:19:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:36.967 15:19:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:37.227 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:14:37.227 15:19:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:37.227 15:19:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:37.227 [2024-11-27 15:19:05.200202] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:37.227 15:19:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:37.227 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:37.227 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:37.227 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:37.227 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:37.227 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:37.227 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:37.227 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:37.227 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:37.227 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:37.227 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:37.227 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:37.227 15:19:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:37.227 15:19:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:37.227 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:37.228 15:19:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:37.228 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:37.228 "name": "Existed_Raid", 00:14:37.228 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:37.228 "strip_size_kb": 64, 00:14:37.228 "state": "configuring", 00:14:37.228 "raid_level": "raid5f", 00:14:37.228 "superblock": false, 00:14:37.228 "num_base_bdevs": 4, 00:14:37.228 "num_base_bdevs_discovered": 2, 00:14:37.228 "num_base_bdevs_operational": 4, 00:14:37.228 "base_bdevs_list": [ 00:14:37.228 { 00:14:37.228 "name": "BaseBdev1", 00:14:37.228 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:37.228 "is_configured": false, 00:14:37.228 "data_offset": 0, 00:14:37.228 "data_size": 0 00:14:37.228 }, 00:14:37.228 { 00:14:37.228 "name": null, 00:14:37.228 "uuid": "eb742695-1fb8-4278-affc-c00aeeda44e3", 00:14:37.228 "is_configured": false, 00:14:37.228 "data_offset": 0, 00:14:37.228 "data_size": 65536 00:14:37.228 }, 00:14:37.228 { 00:14:37.228 "name": "BaseBdev3", 00:14:37.228 "uuid": "ffd4df99-7b13-4687-bcb5-b3ec1525f886", 00:14:37.228 "is_configured": true, 00:14:37.228 "data_offset": 0, 00:14:37.228 "data_size": 65536 00:14:37.228 }, 00:14:37.228 { 00:14:37.228 "name": "BaseBdev4", 00:14:37.228 "uuid": "3c7bbd24-5ff0-4c4b-9ede-e4f9a28a1ec0", 00:14:37.228 "is_configured": true, 00:14:37.228 "data_offset": 0, 00:14:37.228 "data_size": 65536 00:14:37.228 } 00:14:37.228 ] 00:14:37.228 }' 00:14:37.228 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:37.228 15:19:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:37.796 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:37.796 15:19:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:37.796 15:19:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:37.796 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:37.796 15:19:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:37.796 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:14:37.796 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:37.796 15:19:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:37.796 15:19:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:37.796 [2024-11-27 15:19:05.728530] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:37.796 BaseBdev1 00:14:37.796 15:19:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:37.796 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:14:37.796 15:19:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:14:37.796 15:19:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:37.796 15:19:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:37.796 15:19:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:37.796 15:19:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:37.796 15:19:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:37.796 15:19:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:37.796 15:19:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:37.796 15:19:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:37.796 15:19:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:37.796 15:19:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:37.796 15:19:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:37.796 [ 00:14:37.796 { 00:14:37.796 "name": "BaseBdev1", 00:14:37.796 "aliases": [ 00:14:37.796 "c8db558f-883f-4bbb-9eae-93f74d9173e9" 00:14:37.796 ], 00:14:37.796 "product_name": "Malloc disk", 00:14:37.796 "block_size": 512, 00:14:37.796 "num_blocks": 65536, 00:14:37.796 "uuid": "c8db558f-883f-4bbb-9eae-93f74d9173e9", 00:14:37.796 "assigned_rate_limits": { 00:14:37.796 "rw_ios_per_sec": 0, 00:14:37.796 "rw_mbytes_per_sec": 0, 00:14:37.796 "r_mbytes_per_sec": 0, 00:14:37.796 "w_mbytes_per_sec": 0 00:14:37.796 }, 00:14:37.796 "claimed": true, 00:14:37.796 "claim_type": "exclusive_write", 00:14:37.796 "zoned": false, 00:14:37.796 "supported_io_types": { 00:14:37.796 "read": true, 00:14:37.796 "write": true, 00:14:37.796 "unmap": true, 00:14:37.796 "flush": true, 00:14:37.796 "reset": true, 00:14:37.796 "nvme_admin": false, 00:14:37.796 "nvme_io": false, 00:14:37.796 "nvme_io_md": false, 00:14:37.796 "write_zeroes": true, 00:14:37.796 "zcopy": true, 00:14:37.796 "get_zone_info": false, 00:14:37.796 "zone_management": false, 00:14:37.796 "zone_append": false, 00:14:37.796 "compare": false, 00:14:37.796 "compare_and_write": false, 00:14:37.796 "abort": true, 00:14:37.796 "seek_hole": false, 00:14:37.796 "seek_data": false, 00:14:37.796 "copy": true, 00:14:37.796 "nvme_iov_md": false 00:14:37.796 }, 00:14:37.796 "memory_domains": [ 00:14:37.796 { 00:14:37.796 "dma_device_id": "system", 00:14:37.796 "dma_device_type": 1 00:14:37.796 }, 00:14:37.796 { 00:14:37.796 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:37.796 "dma_device_type": 2 00:14:37.796 } 00:14:37.796 ], 00:14:37.796 "driver_specific": {} 00:14:37.796 } 00:14:37.796 ] 00:14:37.796 15:19:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:37.796 15:19:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:37.796 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:37.796 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:37.796 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:37.796 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:37.797 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:37.797 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:37.797 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:37.797 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:37.797 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:37.797 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:37.797 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:37.797 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:37.797 15:19:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:37.797 15:19:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:37.797 15:19:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:37.797 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:37.797 "name": "Existed_Raid", 00:14:37.797 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:37.797 "strip_size_kb": 64, 00:14:37.797 "state": "configuring", 00:14:37.797 "raid_level": "raid5f", 00:14:37.797 "superblock": false, 00:14:37.797 "num_base_bdevs": 4, 00:14:37.797 "num_base_bdevs_discovered": 3, 00:14:37.797 "num_base_bdevs_operational": 4, 00:14:37.797 "base_bdevs_list": [ 00:14:37.797 { 00:14:37.797 "name": "BaseBdev1", 00:14:37.797 "uuid": "c8db558f-883f-4bbb-9eae-93f74d9173e9", 00:14:37.797 "is_configured": true, 00:14:37.797 "data_offset": 0, 00:14:37.797 "data_size": 65536 00:14:37.797 }, 00:14:37.797 { 00:14:37.797 "name": null, 00:14:37.797 "uuid": "eb742695-1fb8-4278-affc-c00aeeda44e3", 00:14:37.797 "is_configured": false, 00:14:37.797 "data_offset": 0, 00:14:37.797 "data_size": 65536 00:14:37.797 }, 00:14:37.797 { 00:14:37.797 "name": "BaseBdev3", 00:14:37.797 "uuid": "ffd4df99-7b13-4687-bcb5-b3ec1525f886", 00:14:37.797 "is_configured": true, 00:14:37.797 "data_offset": 0, 00:14:37.797 "data_size": 65536 00:14:37.797 }, 00:14:37.797 { 00:14:37.797 "name": "BaseBdev4", 00:14:37.797 "uuid": "3c7bbd24-5ff0-4c4b-9ede-e4f9a28a1ec0", 00:14:37.797 "is_configured": true, 00:14:37.797 "data_offset": 0, 00:14:37.797 "data_size": 65536 00:14:37.797 } 00:14:37.797 ] 00:14:37.797 }' 00:14:37.797 15:19:05 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:37.797 15:19:05 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:38.366 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:38.366 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:38.366 15:19:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:38.366 15:19:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:38.366 15:19:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:38.366 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:14:38.366 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:14:38.366 15:19:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:38.366 15:19:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:38.366 [2024-11-27 15:19:06.247729] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:38.366 15:19:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:38.366 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:38.366 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:38.366 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:38.366 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:38.366 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:38.366 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:38.366 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:38.366 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:38.366 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:38.366 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:38.366 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:38.366 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:38.366 15:19:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:38.366 15:19:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:38.366 15:19:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:38.366 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:38.366 "name": "Existed_Raid", 00:14:38.366 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:38.366 "strip_size_kb": 64, 00:14:38.366 "state": "configuring", 00:14:38.366 "raid_level": "raid5f", 00:14:38.366 "superblock": false, 00:14:38.366 "num_base_bdevs": 4, 00:14:38.366 "num_base_bdevs_discovered": 2, 00:14:38.366 "num_base_bdevs_operational": 4, 00:14:38.366 "base_bdevs_list": [ 00:14:38.366 { 00:14:38.366 "name": "BaseBdev1", 00:14:38.366 "uuid": "c8db558f-883f-4bbb-9eae-93f74d9173e9", 00:14:38.366 "is_configured": true, 00:14:38.366 "data_offset": 0, 00:14:38.366 "data_size": 65536 00:14:38.366 }, 00:14:38.366 { 00:14:38.366 "name": null, 00:14:38.366 "uuid": "eb742695-1fb8-4278-affc-c00aeeda44e3", 00:14:38.366 "is_configured": false, 00:14:38.366 "data_offset": 0, 00:14:38.366 "data_size": 65536 00:14:38.366 }, 00:14:38.366 { 00:14:38.366 "name": null, 00:14:38.366 "uuid": "ffd4df99-7b13-4687-bcb5-b3ec1525f886", 00:14:38.366 "is_configured": false, 00:14:38.366 "data_offset": 0, 00:14:38.366 "data_size": 65536 00:14:38.366 }, 00:14:38.366 { 00:14:38.366 "name": "BaseBdev4", 00:14:38.366 "uuid": "3c7bbd24-5ff0-4c4b-9ede-e4f9a28a1ec0", 00:14:38.366 "is_configured": true, 00:14:38.366 "data_offset": 0, 00:14:38.366 "data_size": 65536 00:14:38.366 } 00:14:38.366 ] 00:14:38.366 }' 00:14:38.366 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:38.366 15:19:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:38.627 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:38.627 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:38.627 15:19:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:38.627 15:19:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:38.888 15:19:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:38.888 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:14:38.888 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:14:38.888 15:19:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:38.888 15:19:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:38.888 [2024-11-27 15:19:06.758973] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:38.888 15:19:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:38.888 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:38.888 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:38.888 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:38.888 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:38.888 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:38.888 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:38.888 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:38.888 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:38.888 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:38.888 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:38.888 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:38.888 15:19:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:38.888 15:19:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:38.888 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:38.888 15:19:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:38.888 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:38.888 "name": "Existed_Raid", 00:14:38.888 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:38.888 "strip_size_kb": 64, 00:14:38.888 "state": "configuring", 00:14:38.888 "raid_level": "raid5f", 00:14:38.888 "superblock": false, 00:14:38.888 "num_base_bdevs": 4, 00:14:38.888 "num_base_bdevs_discovered": 3, 00:14:38.888 "num_base_bdevs_operational": 4, 00:14:38.888 "base_bdevs_list": [ 00:14:38.888 { 00:14:38.888 "name": "BaseBdev1", 00:14:38.888 "uuid": "c8db558f-883f-4bbb-9eae-93f74d9173e9", 00:14:38.888 "is_configured": true, 00:14:38.888 "data_offset": 0, 00:14:38.888 "data_size": 65536 00:14:38.888 }, 00:14:38.888 { 00:14:38.888 "name": null, 00:14:38.888 "uuid": "eb742695-1fb8-4278-affc-c00aeeda44e3", 00:14:38.888 "is_configured": false, 00:14:38.888 "data_offset": 0, 00:14:38.888 "data_size": 65536 00:14:38.888 }, 00:14:38.888 { 00:14:38.888 "name": "BaseBdev3", 00:14:38.888 "uuid": "ffd4df99-7b13-4687-bcb5-b3ec1525f886", 00:14:38.888 "is_configured": true, 00:14:38.888 "data_offset": 0, 00:14:38.888 "data_size": 65536 00:14:38.888 }, 00:14:38.888 { 00:14:38.888 "name": "BaseBdev4", 00:14:38.888 "uuid": "3c7bbd24-5ff0-4c4b-9ede-e4f9a28a1ec0", 00:14:38.888 "is_configured": true, 00:14:38.888 "data_offset": 0, 00:14:38.888 "data_size": 65536 00:14:38.888 } 00:14:38.888 ] 00:14:38.888 }' 00:14:38.888 15:19:06 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:38.888 15:19:06 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:39.148 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:39.148 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:39.148 15:19:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:39.148 15:19:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:39.408 15:19:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:39.408 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:14:39.408 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:39.408 15:19:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:39.408 15:19:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:39.408 [2024-11-27 15:19:07.294053] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:39.408 15:19:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:39.408 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:39.408 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:39.408 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:39.408 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:39.408 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:39.408 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:39.408 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:39.408 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:39.408 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:39.408 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:39.408 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:39.408 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:39.408 15:19:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:39.408 15:19:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:39.408 15:19:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:39.408 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:39.408 "name": "Existed_Raid", 00:14:39.408 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:39.408 "strip_size_kb": 64, 00:14:39.408 "state": "configuring", 00:14:39.408 "raid_level": "raid5f", 00:14:39.408 "superblock": false, 00:14:39.408 "num_base_bdevs": 4, 00:14:39.408 "num_base_bdevs_discovered": 2, 00:14:39.408 "num_base_bdevs_operational": 4, 00:14:39.408 "base_bdevs_list": [ 00:14:39.408 { 00:14:39.408 "name": null, 00:14:39.408 "uuid": "c8db558f-883f-4bbb-9eae-93f74d9173e9", 00:14:39.408 "is_configured": false, 00:14:39.408 "data_offset": 0, 00:14:39.408 "data_size": 65536 00:14:39.408 }, 00:14:39.408 { 00:14:39.408 "name": null, 00:14:39.408 "uuid": "eb742695-1fb8-4278-affc-c00aeeda44e3", 00:14:39.408 "is_configured": false, 00:14:39.408 "data_offset": 0, 00:14:39.408 "data_size": 65536 00:14:39.408 }, 00:14:39.408 { 00:14:39.408 "name": "BaseBdev3", 00:14:39.408 "uuid": "ffd4df99-7b13-4687-bcb5-b3ec1525f886", 00:14:39.408 "is_configured": true, 00:14:39.408 "data_offset": 0, 00:14:39.408 "data_size": 65536 00:14:39.408 }, 00:14:39.408 { 00:14:39.408 "name": "BaseBdev4", 00:14:39.408 "uuid": "3c7bbd24-5ff0-4c4b-9ede-e4f9a28a1ec0", 00:14:39.408 "is_configured": true, 00:14:39.408 "data_offset": 0, 00:14:39.408 "data_size": 65536 00:14:39.408 } 00:14:39.408 ] 00:14:39.408 }' 00:14:39.408 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:39.408 15:19:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:39.667 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:39.667 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:39.667 15:19:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:39.667 15:19:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:39.667 15:19:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:39.667 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:14:39.668 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:14:39.668 15:19:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:39.668 15:19:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:39.668 [2024-11-27 15:19:07.769791] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:39.927 15:19:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:39.927 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:39.927 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:39.927 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:39.927 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:39.927 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:39.927 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:39.927 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:39.927 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:39.927 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:39.927 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:39.927 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:39.927 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:39.927 15:19:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:39.927 15:19:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:39.927 15:19:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:39.927 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:39.927 "name": "Existed_Raid", 00:14:39.927 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:39.927 "strip_size_kb": 64, 00:14:39.927 "state": "configuring", 00:14:39.927 "raid_level": "raid5f", 00:14:39.927 "superblock": false, 00:14:39.927 "num_base_bdevs": 4, 00:14:39.927 "num_base_bdevs_discovered": 3, 00:14:39.927 "num_base_bdevs_operational": 4, 00:14:39.927 "base_bdevs_list": [ 00:14:39.927 { 00:14:39.927 "name": null, 00:14:39.927 "uuid": "c8db558f-883f-4bbb-9eae-93f74d9173e9", 00:14:39.928 "is_configured": false, 00:14:39.928 "data_offset": 0, 00:14:39.928 "data_size": 65536 00:14:39.928 }, 00:14:39.928 { 00:14:39.928 "name": "BaseBdev2", 00:14:39.928 "uuid": "eb742695-1fb8-4278-affc-c00aeeda44e3", 00:14:39.928 "is_configured": true, 00:14:39.928 "data_offset": 0, 00:14:39.928 "data_size": 65536 00:14:39.928 }, 00:14:39.928 { 00:14:39.928 "name": "BaseBdev3", 00:14:39.928 "uuid": "ffd4df99-7b13-4687-bcb5-b3ec1525f886", 00:14:39.928 "is_configured": true, 00:14:39.928 "data_offset": 0, 00:14:39.928 "data_size": 65536 00:14:39.928 }, 00:14:39.928 { 00:14:39.928 "name": "BaseBdev4", 00:14:39.928 "uuid": "3c7bbd24-5ff0-4c4b-9ede-e4f9a28a1ec0", 00:14:39.928 "is_configured": true, 00:14:39.928 "data_offset": 0, 00:14:39.928 "data_size": 65536 00:14:39.928 } 00:14:39.928 ] 00:14:39.928 }' 00:14:39.928 15:19:07 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:39.928 15:19:07 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:40.187 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:40.187 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.187 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:40.187 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:40.187 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.187 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:14:40.187 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:14:40.187 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:40.187 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.187 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u c8db558f-883f-4bbb-9eae-93f74d9173e9 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:40.448 [2024-11-27 15:19:08.321829] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:14:40.448 [2024-11-27 15:19:08.322016] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:14:40.448 [2024-11-27 15:19:08.322052] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:14:40.448 [2024-11-27 15:19:08.322440] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:14:40.448 [2024-11-27 15:19:08.322993] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:14:40.448 [2024-11-27 15:19:08.323053] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006d00 00:14:40.448 [2024-11-27 15:19:08.323312] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:40.448 NewBaseBdev 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:40.448 [ 00:14:40.448 { 00:14:40.448 "name": "NewBaseBdev", 00:14:40.448 "aliases": [ 00:14:40.448 "c8db558f-883f-4bbb-9eae-93f74d9173e9" 00:14:40.448 ], 00:14:40.448 "product_name": "Malloc disk", 00:14:40.448 "block_size": 512, 00:14:40.448 "num_blocks": 65536, 00:14:40.448 "uuid": "c8db558f-883f-4bbb-9eae-93f74d9173e9", 00:14:40.448 "assigned_rate_limits": { 00:14:40.448 "rw_ios_per_sec": 0, 00:14:40.448 "rw_mbytes_per_sec": 0, 00:14:40.448 "r_mbytes_per_sec": 0, 00:14:40.448 "w_mbytes_per_sec": 0 00:14:40.448 }, 00:14:40.448 "claimed": true, 00:14:40.448 "claim_type": "exclusive_write", 00:14:40.448 "zoned": false, 00:14:40.448 "supported_io_types": { 00:14:40.448 "read": true, 00:14:40.448 "write": true, 00:14:40.448 "unmap": true, 00:14:40.448 "flush": true, 00:14:40.448 "reset": true, 00:14:40.448 "nvme_admin": false, 00:14:40.448 "nvme_io": false, 00:14:40.448 "nvme_io_md": false, 00:14:40.448 "write_zeroes": true, 00:14:40.448 "zcopy": true, 00:14:40.448 "get_zone_info": false, 00:14:40.448 "zone_management": false, 00:14:40.448 "zone_append": false, 00:14:40.448 "compare": false, 00:14:40.448 "compare_and_write": false, 00:14:40.448 "abort": true, 00:14:40.448 "seek_hole": false, 00:14:40.448 "seek_data": false, 00:14:40.448 "copy": true, 00:14:40.448 "nvme_iov_md": false 00:14:40.448 }, 00:14:40.448 "memory_domains": [ 00:14:40.448 { 00:14:40.448 "dma_device_id": "system", 00:14:40.448 "dma_device_type": 1 00:14:40.448 }, 00:14:40.448 { 00:14:40.448 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:40.448 "dma_device_type": 2 00:14:40.448 } 00:14:40.448 ], 00:14:40.448 "driver_specific": {} 00:14:40.448 } 00:14:40.448 ] 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.448 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:40.448 "name": "Existed_Raid", 00:14:40.448 "uuid": "57fc1530-1f0e-40cd-aada-d88c816e4d37", 00:14:40.448 "strip_size_kb": 64, 00:14:40.448 "state": "online", 00:14:40.449 "raid_level": "raid5f", 00:14:40.449 "superblock": false, 00:14:40.449 "num_base_bdevs": 4, 00:14:40.449 "num_base_bdevs_discovered": 4, 00:14:40.449 "num_base_bdevs_operational": 4, 00:14:40.449 "base_bdevs_list": [ 00:14:40.449 { 00:14:40.449 "name": "NewBaseBdev", 00:14:40.449 "uuid": "c8db558f-883f-4bbb-9eae-93f74d9173e9", 00:14:40.449 "is_configured": true, 00:14:40.449 "data_offset": 0, 00:14:40.449 "data_size": 65536 00:14:40.449 }, 00:14:40.449 { 00:14:40.449 "name": "BaseBdev2", 00:14:40.449 "uuid": "eb742695-1fb8-4278-affc-c00aeeda44e3", 00:14:40.449 "is_configured": true, 00:14:40.449 "data_offset": 0, 00:14:40.449 "data_size": 65536 00:14:40.449 }, 00:14:40.449 { 00:14:40.449 "name": "BaseBdev3", 00:14:40.449 "uuid": "ffd4df99-7b13-4687-bcb5-b3ec1525f886", 00:14:40.449 "is_configured": true, 00:14:40.449 "data_offset": 0, 00:14:40.449 "data_size": 65536 00:14:40.449 }, 00:14:40.449 { 00:14:40.449 "name": "BaseBdev4", 00:14:40.449 "uuid": "3c7bbd24-5ff0-4c4b-9ede-e4f9a28a1ec0", 00:14:40.449 "is_configured": true, 00:14:40.449 "data_offset": 0, 00:14:40.449 "data_size": 65536 00:14:40.449 } 00:14:40.449 ] 00:14:40.449 }' 00:14:40.449 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:40.449 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:40.709 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:14:40.709 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:40.709 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:40.709 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:40.709 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:40.709 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:40.709 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:40.709 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:40.709 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.709 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:40.709 [2024-11-27 15:19:08.789501] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:40.709 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.969 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:40.969 "name": "Existed_Raid", 00:14:40.969 "aliases": [ 00:14:40.969 "57fc1530-1f0e-40cd-aada-d88c816e4d37" 00:14:40.969 ], 00:14:40.969 "product_name": "Raid Volume", 00:14:40.969 "block_size": 512, 00:14:40.969 "num_blocks": 196608, 00:14:40.969 "uuid": "57fc1530-1f0e-40cd-aada-d88c816e4d37", 00:14:40.969 "assigned_rate_limits": { 00:14:40.969 "rw_ios_per_sec": 0, 00:14:40.969 "rw_mbytes_per_sec": 0, 00:14:40.969 "r_mbytes_per_sec": 0, 00:14:40.969 "w_mbytes_per_sec": 0 00:14:40.969 }, 00:14:40.969 "claimed": false, 00:14:40.969 "zoned": false, 00:14:40.969 "supported_io_types": { 00:14:40.969 "read": true, 00:14:40.969 "write": true, 00:14:40.969 "unmap": false, 00:14:40.969 "flush": false, 00:14:40.969 "reset": true, 00:14:40.969 "nvme_admin": false, 00:14:40.969 "nvme_io": false, 00:14:40.969 "nvme_io_md": false, 00:14:40.969 "write_zeroes": true, 00:14:40.969 "zcopy": false, 00:14:40.969 "get_zone_info": false, 00:14:40.969 "zone_management": false, 00:14:40.969 "zone_append": false, 00:14:40.969 "compare": false, 00:14:40.969 "compare_and_write": false, 00:14:40.969 "abort": false, 00:14:40.969 "seek_hole": false, 00:14:40.969 "seek_data": false, 00:14:40.969 "copy": false, 00:14:40.969 "nvme_iov_md": false 00:14:40.969 }, 00:14:40.969 "driver_specific": { 00:14:40.969 "raid": { 00:14:40.969 "uuid": "57fc1530-1f0e-40cd-aada-d88c816e4d37", 00:14:40.970 "strip_size_kb": 64, 00:14:40.970 "state": "online", 00:14:40.970 "raid_level": "raid5f", 00:14:40.970 "superblock": false, 00:14:40.970 "num_base_bdevs": 4, 00:14:40.970 "num_base_bdevs_discovered": 4, 00:14:40.970 "num_base_bdevs_operational": 4, 00:14:40.970 "base_bdevs_list": [ 00:14:40.970 { 00:14:40.970 "name": "NewBaseBdev", 00:14:40.970 "uuid": "c8db558f-883f-4bbb-9eae-93f74d9173e9", 00:14:40.970 "is_configured": true, 00:14:40.970 "data_offset": 0, 00:14:40.970 "data_size": 65536 00:14:40.970 }, 00:14:40.970 { 00:14:40.970 "name": "BaseBdev2", 00:14:40.970 "uuid": "eb742695-1fb8-4278-affc-c00aeeda44e3", 00:14:40.970 "is_configured": true, 00:14:40.970 "data_offset": 0, 00:14:40.970 "data_size": 65536 00:14:40.970 }, 00:14:40.970 { 00:14:40.970 "name": "BaseBdev3", 00:14:40.970 "uuid": "ffd4df99-7b13-4687-bcb5-b3ec1525f886", 00:14:40.970 "is_configured": true, 00:14:40.970 "data_offset": 0, 00:14:40.970 "data_size": 65536 00:14:40.970 }, 00:14:40.970 { 00:14:40.970 "name": "BaseBdev4", 00:14:40.970 "uuid": "3c7bbd24-5ff0-4c4b-9ede-e4f9a28a1ec0", 00:14:40.970 "is_configured": true, 00:14:40.970 "data_offset": 0, 00:14:40.970 "data_size": 65536 00:14:40.970 } 00:14:40.970 ] 00:14:40.970 } 00:14:40.970 } 00:14:40.970 }' 00:14:40.970 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:40.970 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:14:40.970 BaseBdev2 00:14:40.970 BaseBdev3 00:14:40.970 BaseBdev4' 00:14:40.970 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:40.970 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:40.970 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:40.970 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:40.970 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:14:40.970 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.970 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:40.970 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.970 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:40.970 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:40.970 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:40.970 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:40.970 15:19:08 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:40.970 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.970 15:19:08 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:40.970 15:19:09 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.970 15:19:09 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:40.970 15:19:09 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:40.970 15:19:09 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:40.970 15:19:09 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:40.970 15:19:09 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:40.970 15:19:09 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.970 15:19:09 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:40.970 15:19:09 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:41.230 15:19:09 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:41.230 15:19:09 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:41.230 15:19:09 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:41.230 15:19:09 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:14:41.230 15:19:09 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:41.230 15:19:09 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:41.230 15:19:09 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:41.230 15:19:09 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:41.230 15:19:09 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:41.230 15:19:09 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:41.230 15:19:09 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:41.230 15:19:09 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:41.230 15:19:09 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:41.230 [2024-11-27 15:19:09.136755] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:41.230 [2024-11-27 15:19:09.136838] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:41.230 [2024-11-27 15:19:09.136954] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:41.230 [2024-11-27 15:19:09.137255] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:41.230 [2024-11-27 15:19:09.137315] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name Existed_Raid, state offline 00:14:41.230 15:19:09 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:41.230 15:19:09 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 93430 00:14:41.230 15:19:09 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 93430 ']' 00:14:41.230 15:19:09 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@958 -- # kill -0 93430 00:14:41.230 15:19:09 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@959 -- # uname 00:14:41.230 15:19:09 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:41.230 15:19:09 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 93430 00:14:41.230 15:19:09 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:14:41.230 15:19:09 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:14:41.230 15:19:09 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 93430' 00:14:41.230 killing process with pid 93430 00:14:41.230 15:19:09 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@973 -- # kill 93430 00:14:41.230 [2024-11-27 15:19:09.188760] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:41.230 15:19:09 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@978 -- # wait 93430 00:14:41.230 [2024-11-27 15:19:09.267995] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:14:41.799 00:14:41.799 real 0m9.944s 00:14:41.799 user 0m16.690s 00:14:41.799 sys 0m2.145s 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:41.799 ************************************ 00:14:41.799 END TEST raid5f_state_function_test 00:14:41.799 ************************************ 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:41.799 15:19:09 bdev_raid -- bdev/bdev_raid.sh@987 -- # run_test raid5f_state_function_test_sb raid_state_function_test raid5f 4 true 00:14:41.799 15:19:09 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:14:41.799 15:19:09 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:41.799 15:19:09 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:41.799 ************************************ 00:14:41.799 START TEST raid5f_state_function_test_sb 00:14:41.799 ************************************ 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid5f 4 true 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:14:41.799 Process raid pid: 94085 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=94085 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 94085' 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 94085 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 94085 ']' 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:41.799 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:41.799 15:19:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:41.799 [2024-11-27 15:19:09.791780] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:14:41.799 [2024-11-27 15:19:09.792013] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:42.059 [2024-11-27 15:19:09.965004] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:42.059 [2024-11-27 15:19:10.004490] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:42.059 [2024-11-27 15:19:10.081682] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:42.059 [2024-11-27 15:19:10.081820] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:42.630 15:19:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:42.630 15:19:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:14:42.630 15:19:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:42.630 15:19:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.630 15:19:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:42.630 [2024-11-27 15:19:10.596598] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:42.630 [2024-11-27 15:19:10.596743] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:42.630 [2024-11-27 15:19:10.596777] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:42.630 [2024-11-27 15:19:10.596806] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:42.630 [2024-11-27 15:19:10.596826] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:42.630 [2024-11-27 15:19:10.596856] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:42.630 [2024-11-27 15:19:10.596877] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:42.630 [2024-11-27 15:19:10.596950] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:42.630 15:19:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.630 15:19:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:42.630 15:19:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:42.630 15:19:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:42.630 15:19:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:42.630 15:19:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:42.630 15:19:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:42.630 15:19:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:42.630 15:19:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:42.630 15:19:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:42.630 15:19:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:42.630 15:19:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:42.630 15:19:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:42.630 15:19:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.630 15:19:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:42.630 15:19:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.630 15:19:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:42.630 "name": "Existed_Raid", 00:14:42.630 "uuid": "b12db91e-bc34-48c9-adb4-f300687da8ce", 00:14:42.630 "strip_size_kb": 64, 00:14:42.630 "state": "configuring", 00:14:42.630 "raid_level": "raid5f", 00:14:42.630 "superblock": true, 00:14:42.630 "num_base_bdevs": 4, 00:14:42.630 "num_base_bdevs_discovered": 0, 00:14:42.630 "num_base_bdevs_operational": 4, 00:14:42.630 "base_bdevs_list": [ 00:14:42.630 { 00:14:42.630 "name": "BaseBdev1", 00:14:42.630 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:42.630 "is_configured": false, 00:14:42.630 "data_offset": 0, 00:14:42.630 "data_size": 0 00:14:42.630 }, 00:14:42.630 { 00:14:42.630 "name": "BaseBdev2", 00:14:42.630 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:42.630 "is_configured": false, 00:14:42.630 "data_offset": 0, 00:14:42.630 "data_size": 0 00:14:42.630 }, 00:14:42.630 { 00:14:42.630 "name": "BaseBdev3", 00:14:42.630 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:42.630 "is_configured": false, 00:14:42.630 "data_offset": 0, 00:14:42.630 "data_size": 0 00:14:42.630 }, 00:14:42.630 { 00:14:42.630 "name": "BaseBdev4", 00:14:42.630 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:42.630 "is_configured": false, 00:14:42.630 "data_offset": 0, 00:14:42.631 "data_size": 0 00:14:42.631 } 00:14:42.631 ] 00:14:42.631 }' 00:14:42.631 15:19:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:42.631 15:19:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:43.200 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:43.200 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.200 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:43.200 [2024-11-27 15:19:11.051830] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:43.200 [2024-11-27 15:19:11.051948] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:14:43.200 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.200 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:43.200 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.200 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:43.200 [2024-11-27 15:19:11.063833] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:43.200 [2024-11-27 15:19:11.063939] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:43.201 [2024-11-27 15:19:11.063971] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:43.201 [2024-11-27 15:19:11.063999] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:43.201 [2024-11-27 15:19:11.064021] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:43.201 [2024-11-27 15:19:11.064047] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:43.201 [2024-11-27 15:19:11.064069] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:43.201 [2024-11-27 15:19:11.064096] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:43.201 [2024-11-27 15:19:11.091236] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:43.201 BaseBdev1 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:43.201 [ 00:14:43.201 { 00:14:43.201 "name": "BaseBdev1", 00:14:43.201 "aliases": [ 00:14:43.201 "09867161-191c-4e0e-b301-0517a3615b75" 00:14:43.201 ], 00:14:43.201 "product_name": "Malloc disk", 00:14:43.201 "block_size": 512, 00:14:43.201 "num_blocks": 65536, 00:14:43.201 "uuid": "09867161-191c-4e0e-b301-0517a3615b75", 00:14:43.201 "assigned_rate_limits": { 00:14:43.201 "rw_ios_per_sec": 0, 00:14:43.201 "rw_mbytes_per_sec": 0, 00:14:43.201 "r_mbytes_per_sec": 0, 00:14:43.201 "w_mbytes_per_sec": 0 00:14:43.201 }, 00:14:43.201 "claimed": true, 00:14:43.201 "claim_type": "exclusive_write", 00:14:43.201 "zoned": false, 00:14:43.201 "supported_io_types": { 00:14:43.201 "read": true, 00:14:43.201 "write": true, 00:14:43.201 "unmap": true, 00:14:43.201 "flush": true, 00:14:43.201 "reset": true, 00:14:43.201 "nvme_admin": false, 00:14:43.201 "nvme_io": false, 00:14:43.201 "nvme_io_md": false, 00:14:43.201 "write_zeroes": true, 00:14:43.201 "zcopy": true, 00:14:43.201 "get_zone_info": false, 00:14:43.201 "zone_management": false, 00:14:43.201 "zone_append": false, 00:14:43.201 "compare": false, 00:14:43.201 "compare_and_write": false, 00:14:43.201 "abort": true, 00:14:43.201 "seek_hole": false, 00:14:43.201 "seek_data": false, 00:14:43.201 "copy": true, 00:14:43.201 "nvme_iov_md": false 00:14:43.201 }, 00:14:43.201 "memory_domains": [ 00:14:43.201 { 00:14:43.201 "dma_device_id": "system", 00:14:43.201 "dma_device_type": 1 00:14:43.201 }, 00:14:43.201 { 00:14:43.201 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:43.201 "dma_device_type": 2 00:14:43.201 } 00:14:43.201 ], 00:14:43.201 "driver_specific": {} 00:14:43.201 } 00:14:43.201 ] 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:43.201 "name": "Existed_Raid", 00:14:43.201 "uuid": "a5143be4-96de-4076-a938-a1073fb9ea13", 00:14:43.201 "strip_size_kb": 64, 00:14:43.201 "state": "configuring", 00:14:43.201 "raid_level": "raid5f", 00:14:43.201 "superblock": true, 00:14:43.201 "num_base_bdevs": 4, 00:14:43.201 "num_base_bdevs_discovered": 1, 00:14:43.201 "num_base_bdevs_operational": 4, 00:14:43.201 "base_bdevs_list": [ 00:14:43.201 { 00:14:43.201 "name": "BaseBdev1", 00:14:43.201 "uuid": "09867161-191c-4e0e-b301-0517a3615b75", 00:14:43.201 "is_configured": true, 00:14:43.201 "data_offset": 2048, 00:14:43.201 "data_size": 63488 00:14:43.201 }, 00:14:43.201 { 00:14:43.201 "name": "BaseBdev2", 00:14:43.201 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:43.201 "is_configured": false, 00:14:43.201 "data_offset": 0, 00:14:43.201 "data_size": 0 00:14:43.201 }, 00:14:43.201 { 00:14:43.201 "name": "BaseBdev3", 00:14:43.201 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:43.201 "is_configured": false, 00:14:43.201 "data_offset": 0, 00:14:43.201 "data_size": 0 00:14:43.201 }, 00:14:43.201 { 00:14:43.201 "name": "BaseBdev4", 00:14:43.201 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:43.201 "is_configured": false, 00:14:43.201 "data_offset": 0, 00:14:43.201 "data_size": 0 00:14:43.201 } 00:14:43.201 ] 00:14:43.201 }' 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:43.201 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:43.462 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:43.462 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.462 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:43.462 [2024-11-27 15:19:11.546472] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:43.462 [2024-11-27 15:19:11.546577] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:14:43.462 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.462 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:43.462 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.462 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:43.463 [2024-11-27 15:19:11.558515] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:43.463 [2024-11-27 15:19:11.560796] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:43.463 [2024-11-27 15:19:11.560885] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:43.463 [2024-11-27 15:19:11.560930] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:43.463 [2024-11-27 15:19:11.560956] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:43.463 [2024-11-27 15:19:11.560978] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:43.463 [2024-11-27 15:19:11.561001] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:43.463 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.463 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:14:43.463 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:43.463 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:43.723 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:43.723 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:43.723 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:43.723 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:43.723 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:43.723 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:43.723 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:43.723 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:43.723 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:43.723 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:43.723 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:43.723 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.723 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:43.723 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.723 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:43.723 "name": "Existed_Raid", 00:14:43.723 "uuid": "eeee316c-54ca-4ee0-9189-e4f07b349ab2", 00:14:43.723 "strip_size_kb": 64, 00:14:43.723 "state": "configuring", 00:14:43.723 "raid_level": "raid5f", 00:14:43.723 "superblock": true, 00:14:43.723 "num_base_bdevs": 4, 00:14:43.723 "num_base_bdevs_discovered": 1, 00:14:43.723 "num_base_bdevs_operational": 4, 00:14:43.723 "base_bdevs_list": [ 00:14:43.723 { 00:14:43.723 "name": "BaseBdev1", 00:14:43.723 "uuid": "09867161-191c-4e0e-b301-0517a3615b75", 00:14:43.723 "is_configured": true, 00:14:43.723 "data_offset": 2048, 00:14:43.723 "data_size": 63488 00:14:43.723 }, 00:14:43.723 { 00:14:43.723 "name": "BaseBdev2", 00:14:43.723 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:43.723 "is_configured": false, 00:14:43.723 "data_offset": 0, 00:14:43.723 "data_size": 0 00:14:43.723 }, 00:14:43.723 { 00:14:43.723 "name": "BaseBdev3", 00:14:43.723 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:43.723 "is_configured": false, 00:14:43.723 "data_offset": 0, 00:14:43.723 "data_size": 0 00:14:43.723 }, 00:14:43.723 { 00:14:43.723 "name": "BaseBdev4", 00:14:43.723 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:43.723 "is_configured": false, 00:14:43.723 "data_offset": 0, 00:14:43.723 "data_size": 0 00:14:43.723 } 00:14:43.723 ] 00:14:43.723 }' 00:14:43.723 15:19:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:43.723 15:19:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:43.984 [2024-11-27 15:19:12.030754] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:43.984 BaseBdev2 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:43.984 [ 00:14:43.984 { 00:14:43.984 "name": "BaseBdev2", 00:14:43.984 "aliases": [ 00:14:43.984 "0f1a17d1-a5df-49c0-aeeb-bc10857119d5" 00:14:43.984 ], 00:14:43.984 "product_name": "Malloc disk", 00:14:43.984 "block_size": 512, 00:14:43.984 "num_blocks": 65536, 00:14:43.984 "uuid": "0f1a17d1-a5df-49c0-aeeb-bc10857119d5", 00:14:43.984 "assigned_rate_limits": { 00:14:43.984 "rw_ios_per_sec": 0, 00:14:43.984 "rw_mbytes_per_sec": 0, 00:14:43.984 "r_mbytes_per_sec": 0, 00:14:43.984 "w_mbytes_per_sec": 0 00:14:43.984 }, 00:14:43.984 "claimed": true, 00:14:43.984 "claim_type": "exclusive_write", 00:14:43.984 "zoned": false, 00:14:43.984 "supported_io_types": { 00:14:43.984 "read": true, 00:14:43.984 "write": true, 00:14:43.984 "unmap": true, 00:14:43.984 "flush": true, 00:14:43.984 "reset": true, 00:14:43.984 "nvme_admin": false, 00:14:43.984 "nvme_io": false, 00:14:43.984 "nvme_io_md": false, 00:14:43.984 "write_zeroes": true, 00:14:43.984 "zcopy": true, 00:14:43.984 "get_zone_info": false, 00:14:43.984 "zone_management": false, 00:14:43.984 "zone_append": false, 00:14:43.984 "compare": false, 00:14:43.984 "compare_and_write": false, 00:14:43.984 "abort": true, 00:14:43.984 "seek_hole": false, 00:14:43.984 "seek_data": false, 00:14:43.984 "copy": true, 00:14:43.984 "nvme_iov_md": false 00:14:43.984 }, 00:14:43.984 "memory_domains": [ 00:14:43.984 { 00:14:43.984 "dma_device_id": "system", 00:14:43.984 "dma_device_type": 1 00:14:43.984 }, 00:14:43.984 { 00:14:43.984 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:43.984 "dma_device_type": 2 00:14:43.984 } 00:14:43.984 ], 00:14:43.984 "driver_specific": {} 00:14:43.984 } 00:14:43.984 ] 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:43.984 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:43.985 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.985 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:44.246 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.246 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:44.246 "name": "Existed_Raid", 00:14:44.246 "uuid": "eeee316c-54ca-4ee0-9189-e4f07b349ab2", 00:14:44.246 "strip_size_kb": 64, 00:14:44.246 "state": "configuring", 00:14:44.246 "raid_level": "raid5f", 00:14:44.246 "superblock": true, 00:14:44.246 "num_base_bdevs": 4, 00:14:44.246 "num_base_bdevs_discovered": 2, 00:14:44.246 "num_base_bdevs_operational": 4, 00:14:44.246 "base_bdevs_list": [ 00:14:44.246 { 00:14:44.246 "name": "BaseBdev1", 00:14:44.246 "uuid": "09867161-191c-4e0e-b301-0517a3615b75", 00:14:44.246 "is_configured": true, 00:14:44.246 "data_offset": 2048, 00:14:44.246 "data_size": 63488 00:14:44.246 }, 00:14:44.246 { 00:14:44.246 "name": "BaseBdev2", 00:14:44.246 "uuid": "0f1a17d1-a5df-49c0-aeeb-bc10857119d5", 00:14:44.246 "is_configured": true, 00:14:44.246 "data_offset": 2048, 00:14:44.246 "data_size": 63488 00:14:44.246 }, 00:14:44.246 { 00:14:44.246 "name": "BaseBdev3", 00:14:44.246 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:44.246 "is_configured": false, 00:14:44.246 "data_offset": 0, 00:14:44.246 "data_size": 0 00:14:44.246 }, 00:14:44.246 { 00:14:44.246 "name": "BaseBdev4", 00:14:44.246 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:44.246 "is_configured": false, 00:14:44.246 "data_offset": 0, 00:14:44.246 "data_size": 0 00:14:44.246 } 00:14:44.246 ] 00:14:44.246 }' 00:14:44.246 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:44.246 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:44.507 [2024-11-27 15:19:12.542343] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:44.507 BaseBdev3 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:44.507 [ 00:14:44.507 { 00:14:44.507 "name": "BaseBdev3", 00:14:44.507 "aliases": [ 00:14:44.507 "10378dec-aaf8-4710-8752-c994e109e299" 00:14:44.507 ], 00:14:44.507 "product_name": "Malloc disk", 00:14:44.507 "block_size": 512, 00:14:44.507 "num_blocks": 65536, 00:14:44.507 "uuid": "10378dec-aaf8-4710-8752-c994e109e299", 00:14:44.507 "assigned_rate_limits": { 00:14:44.507 "rw_ios_per_sec": 0, 00:14:44.507 "rw_mbytes_per_sec": 0, 00:14:44.507 "r_mbytes_per_sec": 0, 00:14:44.507 "w_mbytes_per_sec": 0 00:14:44.507 }, 00:14:44.507 "claimed": true, 00:14:44.507 "claim_type": "exclusive_write", 00:14:44.507 "zoned": false, 00:14:44.507 "supported_io_types": { 00:14:44.507 "read": true, 00:14:44.507 "write": true, 00:14:44.507 "unmap": true, 00:14:44.507 "flush": true, 00:14:44.507 "reset": true, 00:14:44.507 "nvme_admin": false, 00:14:44.507 "nvme_io": false, 00:14:44.507 "nvme_io_md": false, 00:14:44.507 "write_zeroes": true, 00:14:44.507 "zcopy": true, 00:14:44.507 "get_zone_info": false, 00:14:44.507 "zone_management": false, 00:14:44.507 "zone_append": false, 00:14:44.507 "compare": false, 00:14:44.507 "compare_and_write": false, 00:14:44.507 "abort": true, 00:14:44.507 "seek_hole": false, 00:14:44.507 "seek_data": false, 00:14:44.507 "copy": true, 00:14:44.507 "nvme_iov_md": false 00:14:44.507 }, 00:14:44.507 "memory_domains": [ 00:14:44.507 { 00:14:44.507 "dma_device_id": "system", 00:14:44.507 "dma_device_type": 1 00:14:44.507 }, 00:14:44.507 { 00:14:44.507 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:44.507 "dma_device_type": 2 00:14:44.507 } 00:14:44.507 ], 00:14:44.507 "driver_specific": {} 00:14:44.507 } 00:14:44.507 ] 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:44.507 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.767 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:44.767 "name": "Existed_Raid", 00:14:44.767 "uuid": "eeee316c-54ca-4ee0-9189-e4f07b349ab2", 00:14:44.767 "strip_size_kb": 64, 00:14:44.767 "state": "configuring", 00:14:44.767 "raid_level": "raid5f", 00:14:44.767 "superblock": true, 00:14:44.767 "num_base_bdevs": 4, 00:14:44.767 "num_base_bdevs_discovered": 3, 00:14:44.767 "num_base_bdevs_operational": 4, 00:14:44.767 "base_bdevs_list": [ 00:14:44.767 { 00:14:44.767 "name": "BaseBdev1", 00:14:44.767 "uuid": "09867161-191c-4e0e-b301-0517a3615b75", 00:14:44.767 "is_configured": true, 00:14:44.767 "data_offset": 2048, 00:14:44.767 "data_size": 63488 00:14:44.767 }, 00:14:44.767 { 00:14:44.767 "name": "BaseBdev2", 00:14:44.767 "uuid": "0f1a17d1-a5df-49c0-aeeb-bc10857119d5", 00:14:44.767 "is_configured": true, 00:14:44.767 "data_offset": 2048, 00:14:44.767 "data_size": 63488 00:14:44.767 }, 00:14:44.767 { 00:14:44.767 "name": "BaseBdev3", 00:14:44.767 "uuid": "10378dec-aaf8-4710-8752-c994e109e299", 00:14:44.767 "is_configured": true, 00:14:44.767 "data_offset": 2048, 00:14:44.767 "data_size": 63488 00:14:44.767 }, 00:14:44.767 { 00:14:44.767 "name": "BaseBdev4", 00:14:44.767 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:44.767 "is_configured": false, 00:14:44.767 "data_offset": 0, 00:14:44.767 "data_size": 0 00:14:44.767 } 00:14:44.767 ] 00:14:44.767 }' 00:14:44.767 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:44.767 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:45.028 15:19:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:14:45.028 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.028 15:19:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:45.028 [2024-11-27 15:19:13.018429] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:45.028 [2024-11-27 15:19:13.018802] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:14:45.028 [2024-11-27 15:19:13.018864] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:45.028 [2024-11-27 15:19:13.019273] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:14:45.028 BaseBdev4 00:14:45.028 [2024-11-27 15:19:13.019928] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:14:45.028 [2024-11-27 15:19:13.019965] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:14:45.029 [2024-11-27 15:19:13.020119] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:45.029 [ 00:14:45.029 { 00:14:45.029 "name": "BaseBdev4", 00:14:45.029 "aliases": [ 00:14:45.029 "38ab3518-d65a-4b75-8695-8bad1a4255ae" 00:14:45.029 ], 00:14:45.029 "product_name": "Malloc disk", 00:14:45.029 "block_size": 512, 00:14:45.029 "num_blocks": 65536, 00:14:45.029 "uuid": "38ab3518-d65a-4b75-8695-8bad1a4255ae", 00:14:45.029 "assigned_rate_limits": { 00:14:45.029 "rw_ios_per_sec": 0, 00:14:45.029 "rw_mbytes_per_sec": 0, 00:14:45.029 "r_mbytes_per_sec": 0, 00:14:45.029 "w_mbytes_per_sec": 0 00:14:45.029 }, 00:14:45.029 "claimed": true, 00:14:45.029 "claim_type": "exclusive_write", 00:14:45.029 "zoned": false, 00:14:45.029 "supported_io_types": { 00:14:45.029 "read": true, 00:14:45.029 "write": true, 00:14:45.029 "unmap": true, 00:14:45.029 "flush": true, 00:14:45.029 "reset": true, 00:14:45.029 "nvme_admin": false, 00:14:45.029 "nvme_io": false, 00:14:45.029 "nvme_io_md": false, 00:14:45.029 "write_zeroes": true, 00:14:45.029 "zcopy": true, 00:14:45.029 "get_zone_info": false, 00:14:45.029 "zone_management": false, 00:14:45.029 "zone_append": false, 00:14:45.029 "compare": false, 00:14:45.029 "compare_and_write": false, 00:14:45.029 "abort": true, 00:14:45.029 "seek_hole": false, 00:14:45.029 "seek_data": false, 00:14:45.029 "copy": true, 00:14:45.029 "nvme_iov_md": false 00:14:45.029 }, 00:14:45.029 "memory_domains": [ 00:14:45.029 { 00:14:45.029 "dma_device_id": "system", 00:14:45.029 "dma_device_type": 1 00:14:45.029 }, 00:14:45.029 { 00:14:45.029 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:45.029 "dma_device_type": 2 00:14:45.029 } 00:14:45.029 ], 00:14:45.029 "driver_specific": {} 00:14:45.029 } 00:14:45.029 ] 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:45.029 "name": "Existed_Raid", 00:14:45.029 "uuid": "eeee316c-54ca-4ee0-9189-e4f07b349ab2", 00:14:45.029 "strip_size_kb": 64, 00:14:45.029 "state": "online", 00:14:45.029 "raid_level": "raid5f", 00:14:45.029 "superblock": true, 00:14:45.029 "num_base_bdevs": 4, 00:14:45.029 "num_base_bdevs_discovered": 4, 00:14:45.029 "num_base_bdevs_operational": 4, 00:14:45.029 "base_bdevs_list": [ 00:14:45.029 { 00:14:45.029 "name": "BaseBdev1", 00:14:45.029 "uuid": "09867161-191c-4e0e-b301-0517a3615b75", 00:14:45.029 "is_configured": true, 00:14:45.029 "data_offset": 2048, 00:14:45.029 "data_size": 63488 00:14:45.029 }, 00:14:45.029 { 00:14:45.029 "name": "BaseBdev2", 00:14:45.029 "uuid": "0f1a17d1-a5df-49c0-aeeb-bc10857119d5", 00:14:45.029 "is_configured": true, 00:14:45.029 "data_offset": 2048, 00:14:45.029 "data_size": 63488 00:14:45.029 }, 00:14:45.029 { 00:14:45.029 "name": "BaseBdev3", 00:14:45.029 "uuid": "10378dec-aaf8-4710-8752-c994e109e299", 00:14:45.029 "is_configured": true, 00:14:45.029 "data_offset": 2048, 00:14:45.029 "data_size": 63488 00:14:45.029 }, 00:14:45.029 { 00:14:45.029 "name": "BaseBdev4", 00:14:45.029 "uuid": "38ab3518-d65a-4b75-8695-8bad1a4255ae", 00:14:45.029 "is_configured": true, 00:14:45.029 "data_offset": 2048, 00:14:45.029 "data_size": 63488 00:14:45.029 } 00:14:45.029 ] 00:14:45.029 }' 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:45.029 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:45.600 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:14:45.600 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:45.600 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:45.600 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:45.600 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:14:45.600 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:45.600 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:45.600 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:45.600 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.600 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:45.600 [2024-11-27 15:19:13.482479] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:45.600 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.600 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:45.600 "name": "Existed_Raid", 00:14:45.600 "aliases": [ 00:14:45.600 "eeee316c-54ca-4ee0-9189-e4f07b349ab2" 00:14:45.600 ], 00:14:45.600 "product_name": "Raid Volume", 00:14:45.600 "block_size": 512, 00:14:45.600 "num_blocks": 190464, 00:14:45.600 "uuid": "eeee316c-54ca-4ee0-9189-e4f07b349ab2", 00:14:45.600 "assigned_rate_limits": { 00:14:45.600 "rw_ios_per_sec": 0, 00:14:45.600 "rw_mbytes_per_sec": 0, 00:14:45.600 "r_mbytes_per_sec": 0, 00:14:45.600 "w_mbytes_per_sec": 0 00:14:45.600 }, 00:14:45.600 "claimed": false, 00:14:45.600 "zoned": false, 00:14:45.600 "supported_io_types": { 00:14:45.600 "read": true, 00:14:45.600 "write": true, 00:14:45.600 "unmap": false, 00:14:45.600 "flush": false, 00:14:45.600 "reset": true, 00:14:45.600 "nvme_admin": false, 00:14:45.600 "nvme_io": false, 00:14:45.600 "nvme_io_md": false, 00:14:45.600 "write_zeroes": true, 00:14:45.600 "zcopy": false, 00:14:45.600 "get_zone_info": false, 00:14:45.600 "zone_management": false, 00:14:45.600 "zone_append": false, 00:14:45.600 "compare": false, 00:14:45.600 "compare_and_write": false, 00:14:45.600 "abort": false, 00:14:45.600 "seek_hole": false, 00:14:45.600 "seek_data": false, 00:14:45.600 "copy": false, 00:14:45.600 "nvme_iov_md": false 00:14:45.600 }, 00:14:45.600 "driver_specific": { 00:14:45.600 "raid": { 00:14:45.600 "uuid": "eeee316c-54ca-4ee0-9189-e4f07b349ab2", 00:14:45.600 "strip_size_kb": 64, 00:14:45.600 "state": "online", 00:14:45.600 "raid_level": "raid5f", 00:14:45.600 "superblock": true, 00:14:45.600 "num_base_bdevs": 4, 00:14:45.600 "num_base_bdevs_discovered": 4, 00:14:45.600 "num_base_bdevs_operational": 4, 00:14:45.600 "base_bdevs_list": [ 00:14:45.600 { 00:14:45.600 "name": "BaseBdev1", 00:14:45.600 "uuid": "09867161-191c-4e0e-b301-0517a3615b75", 00:14:45.600 "is_configured": true, 00:14:45.600 "data_offset": 2048, 00:14:45.600 "data_size": 63488 00:14:45.600 }, 00:14:45.600 { 00:14:45.600 "name": "BaseBdev2", 00:14:45.600 "uuid": "0f1a17d1-a5df-49c0-aeeb-bc10857119d5", 00:14:45.600 "is_configured": true, 00:14:45.600 "data_offset": 2048, 00:14:45.600 "data_size": 63488 00:14:45.600 }, 00:14:45.600 { 00:14:45.600 "name": "BaseBdev3", 00:14:45.600 "uuid": "10378dec-aaf8-4710-8752-c994e109e299", 00:14:45.600 "is_configured": true, 00:14:45.600 "data_offset": 2048, 00:14:45.600 "data_size": 63488 00:14:45.600 }, 00:14:45.600 { 00:14:45.600 "name": "BaseBdev4", 00:14:45.600 "uuid": "38ab3518-d65a-4b75-8695-8bad1a4255ae", 00:14:45.600 "is_configured": true, 00:14:45.600 "data_offset": 2048, 00:14:45.600 "data_size": 63488 00:14:45.600 } 00:14:45.600 ] 00:14:45.600 } 00:14:45.600 } 00:14:45.600 }' 00:14:45.600 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:45.600 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:14:45.600 BaseBdev2 00:14:45.600 BaseBdev3 00:14:45.600 BaseBdev4' 00:14:45.600 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:45.600 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:45.600 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:45.600 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:14:45.600 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:45.600 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.600 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:45.600 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.600 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:45.600 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:45.600 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:45.600 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:45.600 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.601 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:45.601 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:45.601 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:45.861 [2024-11-27 15:19:13.829775] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.861 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:45.861 "name": "Existed_Raid", 00:14:45.861 "uuid": "eeee316c-54ca-4ee0-9189-e4f07b349ab2", 00:14:45.861 "strip_size_kb": 64, 00:14:45.861 "state": "online", 00:14:45.861 "raid_level": "raid5f", 00:14:45.861 "superblock": true, 00:14:45.861 "num_base_bdevs": 4, 00:14:45.861 "num_base_bdevs_discovered": 3, 00:14:45.861 "num_base_bdevs_operational": 3, 00:14:45.861 "base_bdevs_list": [ 00:14:45.861 { 00:14:45.861 "name": null, 00:14:45.861 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:45.861 "is_configured": false, 00:14:45.861 "data_offset": 0, 00:14:45.861 "data_size": 63488 00:14:45.861 }, 00:14:45.861 { 00:14:45.861 "name": "BaseBdev2", 00:14:45.861 "uuid": "0f1a17d1-a5df-49c0-aeeb-bc10857119d5", 00:14:45.861 "is_configured": true, 00:14:45.861 "data_offset": 2048, 00:14:45.861 "data_size": 63488 00:14:45.861 }, 00:14:45.861 { 00:14:45.861 "name": "BaseBdev3", 00:14:45.861 "uuid": "10378dec-aaf8-4710-8752-c994e109e299", 00:14:45.861 "is_configured": true, 00:14:45.861 "data_offset": 2048, 00:14:45.861 "data_size": 63488 00:14:45.861 }, 00:14:45.861 { 00:14:45.861 "name": "BaseBdev4", 00:14:45.862 "uuid": "38ab3518-d65a-4b75-8695-8bad1a4255ae", 00:14:45.862 "is_configured": true, 00:14:45.862 "data_offset": 2048, 00:14:45.862 "data_size": 63488 00:14:45.862 } 00:14:45.862 ] 00:14:45.862 }' 00:14:45.862 15:19:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:45.862 15:19:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:46.439 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:14:46.439 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:46.439 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:46.439 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:46.439 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.439 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:46.439 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.439 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:46.439 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:46.439 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:14:46.439 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.439 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:46.439 [2024-11-27 15:19:14.322142] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:46.439 [2024-11-27 15:19:14.322377] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:46.439 [2024-11-27 15:19:14.343299] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:46.439 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.439 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:46.439 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:46.440 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:46.440 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:46.440 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.440 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:46.440 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.440 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:46.440 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:46.440 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:14:46.440 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.440 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:46.440 [2024-11-27 15:19:14.403246] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:46.440 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.440 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:46.440 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:46.440 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:46.440 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:46.440 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.440 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:46.440 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.440 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:46.440 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:46.440 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:14:46.440 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.440 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:46.440 [2024-11-27 15:19:14.484070] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:14:46.440 [2024-11-27 15:19:14.484203] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:14:46.440 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.440 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:46.440 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:46.440 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:46.440 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:14:46.440 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.440 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:46.440 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:46.701 BaseBdev2 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:46.701 [ 00:14:46.701 { 00:14:46.701 "name": "BaseBdev2", 00:14:46.701 "aliases": [ 00:14:46.701 "c4357d34-75bf-4bdc-b44a-b72f3409c74a" 00:14:46.701 ], 00:14:46.701 "product_name": "Malloc disk", 00:14:46.701 "block_size": 512, 00:14:46.701 "num_blocks": 65536, 00:14:46.701 "uuid": "c4357d34-75bf-4bdc-b44a-b72f3409c74a", 00:14:46.701 "assigned_rate_limits": { 00:14:46.701 "rw_ios_per_sec": 0, 00:14:46.701 "rw_mbytes_per_sec": 0, 00:14:46.701 "r_mbytes_per_sec": 0, 00:14:46.701 "w_mbytes_per_sec": 0 00:14:46.701 }, 00:14:46.701 "claimed": false, 00:14:46.701 "zoned": false, 00:14:46.701 "supported_io_types": { 00:14:46.701 "read": true, 00:14:46.701 "write": true, 00:14:46.701 "unmap": true, 00:14:46.701 "flush": true, 00:14:46.701 "reset": true, 00:14:46.701 "nvme_admin": false, 00:14:46.701 "nvme_io": false, 00:14:46.701 "nvme_io_md": false, 00:14:46.701 "write_zeroes": true, 00:14:46.701 "zcopy": true, 00:14:46.701 "get_zone_info": false, 00:14:46.701 "zone_management": false, 00:14:46.701 "zone_append": false, 00:14:46.701 "compare": false, 00:14:46.701 "compare_and_write": false, 00:14:46.701 "abort": true, 00:14:46.701 "seek_hole": false, 00:14:46.701 "seek_data": false, 00:14:46.701 "copy": true, 00:14:46.701 "nvme_iov_md": false 00:14:46.701 }, 00:14:46.701 "memory_domains": [ 00:14:46.701 { 00:14:46.701 "dma_device_id": "system", 00:14:46.701 "dma_device_type": 1 00:14:46.701 }, 00:14:46.701 { 00:14:46.701 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:46.701 "dma_device_type": 2 00:14:46.701 } 00:14:46.701 ], 00:14:46.701 "driver_specific": {} 00:14:46.701 } 00:14:46.701 ] 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:46.701 BaseBdev3 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:46.701 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:46.702 [ 00:14:46.702 { 00:14:46.702 "name": "BaseBdev3", 00:14:46.702 "aliases": [ 00:14:46.702 "b232d9cd-feb4-49a2-969c-b9928c1339bf" 00:14:46.702 ], 00:14:46.702 "product_name": "Malloc disk", 00:14:46.702 "block_size": 512, 00:14:46.702 "num_blocks": 65536, 00:14:46.702 "uuid": "b232d9cd-feb4-49a2-969c-b9928c1339bf", 00:14:46.702 "assigned_rate_limits": { 00:14:46.702 "rw_ios_per_sec": 0, 00:14:46.702 "rw_mbytes_per_sec": 0, 00:14:46.702 "r_mbytes_per_sec": 0, 00:14:46.702 "w_mbytes_per_sec": 0 00:14:46.702 }, 00:14:46.702 "claimed": false, 00:14:46.702 "zoned": false, 00:14:46.702 "supported_io_types": { 00:14:46.702 "read": true, 00:14:46.702 "write": true, 00:14:46.702 "unmap": true, 00:14:46.702 "flush": true, 00:14:46.702 "reset": true, 00:14:46.702 "nvme_admin": false, 00:14:46.702 "nvme_io": false, 00:14:46.702 "nvme_io_md": false, 00:14:46.702 "write_zeroes": true, 00:14:46.702 "zcopy": true, 00:14:46.702 "get_zone_info": false, 00:14:46.702 "zone_management": false, 00:14:46.702 "zone_append": false, 00:14:46.702 "compare": false, 00:14:46.702 "compare_and_write": false, 00:14:46.702 "abort": true, 00:14:46.702 "seek_hole": false, 00:14:46.702 "seek_data": false, 00:14:46.702 "copy": true, 00:14:46.702 "nvme_iov_md": false 00:14:46.702 }, 00:14:46.702 "memory_domains": [ 00:14:46.702 { 00:14:46.702 "dma_device_id": "system", 00:14:46.702 "dma_device_type": 1 00:14:46.702 }, 00:14:46.702 { 00:14:46.702 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:46.702 "dma_device_type": 2 00:14:46.702 } 00:14:46.702 ], 00:14:46.702 "driver_specific": {} 00:14:46.702 } 00:14:46.702 ] 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:46.702 BaseBdev4 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:46.702 [ 00:14:46.702 { 00:14:46.702 "name": "BaseBdev4", 00:14:46.702 "aliases": [ 00:14:46.702 "ebaf47b1-0cc0-4ae9-88fb-23d780b39614" 00:14:46.702 ], 00:14:46.702 "product_name": "Malloc disk", 00:14:46.702 "block_size": 512, 00:14:46.702 "num_blocks": 65536, 00:14:46.702 "uuid": "ebaf47b1-0cc0-4ae9-88fb-23d780b39614", 00:14:46.702 "assigned_rate_limits": { 00:14:46.702 "rw_ios_per_sec": 0, 00:14:46.702 "rw_mbytes_per_sec": 0, 00:14:46.702 "r_mbytes_per_sec": 0, 00:14:46.702 "w_mbytes_per_sec": 0 00:14:46.702 }, 00:14:46.702 "claimed": false, 00:14:46.702 "zoned": false, 00:14:46.702 "supported_io_types": { 00:14:46.702 "read": true, 00:14:46.702 "write": true, 00:14:46.702 "unmap": true, 00:14:46.702 "flush": true, 00:14:46.702 "reset": true, 00:14:46.702 "nvme_admin": false, 00:14:46.702 "nvme_io": false, 00:14:46.702 "nvme_io_md": false, 00:14:46.702 "write_zeroes": true, 00:14:46.702 "zcopy": true, 00:14:46.702 "get_zone_info": false, 00:14:46.702 "zone_management": false, 00:14:46.702 "zone_append": false, 00:14:46.702 "compare": false, 00:14:46.702 "compare_and_write": false, 00:14:46.702 "abort": true, 00:14:46.702 "seek_hole": false, 00:14:46.702 "seek_data": false, 00:14:46.702 "copy": true, 00:14:46.702 "nvme_iov_md": false 00:14:46.702 }, 00:14:46.702 "memory_domains": [ 00:14:46.702 { 00:14:46.702 "dma_device_id": "system", 00:14:46.702 "dma_device_type": 1 00:14:46.702 }, 00:14:46.702 { 00:14:46.702 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:46.702 "dma_device_type": 2 00:14:46.702 } 00:14:46.702 ], 00:14:46.702 "driver_specific": {} 00:14:46.702 } 00:14:46.702 ] 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:46.702 [2024-11-27 15:19:14.736245] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:46.702 [2024-11-27 15:19:14.736390] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:46.702 [2024-11-27 15:19:14.736440] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:46.702 [2024-11-27 15:19:14.738581] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:46.702 [2024-11-27 15:19:14.738689] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.702 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:46.702 "name": "Existed_Raid", 00:14:46.702 "uuid": "900e6a7e-077f-474d-8a65-557a0dfacc8e", 00:14:46.702 "strip_size_kb": 64, 00:14:46.702 "state": "configuring", 00:14:46.702 "raid_level": "raid5f", 00:14:46.702 "superblock": true, 00:14:46.702 "num_base_bdevs": 4, 00:14:46.702 "num_base_bdevs_discovered": 3, 00:14:46.702 "num_base_bdevs_operational": 4, 00:14:46.702 "base_bdevs_list": [ 00:14:46.702 { 00:14:46.702 "name": "BaseBdev1", 00:14:46.702 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:46.702 "is_configured": false, 00:14:46.702 "data_offset": 0, 00:14:46.702 "data_size": 0 00:14:46.702 }, 00:14:46.702 { 00:14:46.702 "name": "BaseBdev2", 00:14:46.702 "uuid": "c4357d34-75bf-4bdc-b44a-b72f3409c74a", 00:14:46.702 "is_configured": true, 00:14:46.702 "data_offset": 2048, 00:14:46.702 "data_size": 63488 00:14:46.703 }, 00:14:46.703 { 00:14:46.703 "name": "BaseBdev3", 00:14:46.703 "uuid": "b232d9cd-feb4-49a2-969c-b9928c1339bf", 00:14:46.703 "is_configured": true, 00:14:46.703 "data_offset": 2048, 00:14:46.703 "data_size": 63488 00:14:46.703 }, 00:14:46.703 { 00:14:46.703 "name": "BaseBdev4", 00:14:46.703 "uuid": "ebaf47b1-0cc0-4ae9-88fb-23d780b39614", 00:14:46.703 "is_configured": true, 00:14:46.703 "data_offset": 2048, 00:14:46.703 "data_size": 63488 00:14:46.703 } 00:14:46.703 ] 00:14:46.703 }' 00:14:46.703 15:19:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:46.703 15:19:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:47.273 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:14:47.273 15:19:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.273 15:19:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:47.273 [2024-11-27 15:19:15.207554] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:47.273 15:19:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.273 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:47.273 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:47.273 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:47.273 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:47.273 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:47.273 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:47.273 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:47.273 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:47.273 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:47.273 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:47.273 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:47.273 15:19:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.273 15:19:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:47.273 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:47.273 15:19:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.273 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:47.273 "name": "Existed_Raid", 00:14:47.273 "uuid": "900e6a7e-077f-474d-8a65-557a0dfacc8e", 00:14:47.273 "strip_size_kb": 64, 00:14:47.273 "state": "configuring", 00:14:47.273 "raid_level": "raid5f", 00:14:47.273 "superblock": true, 00:14:47.273 "num_base_bdevs": 4, 00:14:47.273 "num_base_bdevs_discovered": 2, 00:14:47.273 "num_base_bdevs_operational": 4, 00:14:47.273 "base_bdevs_list": [ 00:14:47.273 { 00:14:47.273 "name": "BaseBdev1", 00:14:47.273 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:47.273 "is_configured": false, 00:14:47.273 "data_offset": 0, 00:14:47.273 "data_size": 0 00:14:47.273 }, 00:14:47.273 { 00:14:47.273 "name": null, 00:14:47.273 "uuid": "c4357d34-75bf-4bdc-b44a-b72f3409c74a", 00:14:47.273 "is_configured": false, 00:14:47.273 "data_offset": 0, 00:14:47.273 "data_size": 63488 00:14:47.273 }, 00:14:47.273 { 00:14:47.273 "name": "BaseBdev3", 00:14:47.273 "uuid": "b232d9cd-feb4-49a2-969c-b9928c1339bf", 00:14:47.273 "is_configured": true, 00:14:47.273 "data_offset": 2048, 00:14:47.273 "data_size": 63488 00:14:47.273 }, 00:14:47.273 { 00:14:47.273 "name": "BaseBdev4", 00:14:47.273 "uuid": "ebaf47b1-0cc0-4ae9-88fb-23d780b39614", 00:14:47.273 "is_configured": true, 00:14:47.273 "data_offset": 2048, 00:14:47.273 "data_size": 63488 00:14:47.273 } 00:14:47.273 ] 00:14:47.273 }' 00:14:47.273 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:47.273 15:19:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:47.843 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:47.843 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:47.843 15:19:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.843 15:19:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:47.843 15:19:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.843 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:14:47.843 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:47.843 15:19:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.843 15:19:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:47.843 BaseBdev1 00:14:47.843 [2024-11-27 15:19:15.712022] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:47.843 15:19:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.843 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:14:47.843 15:19:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:14:47.843 15:19:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:47.843 15:19:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:47.843 15:19:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:47.843 15:19:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:47.843 15:19:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:47.844 15:19:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.844 15:19:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:47.844 15:19:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.844 15:19:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:47.844 15:19:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.844 15:19:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:47.844 [ 00:14:47.844 { 00:14:47.844 "name": "BaseBdev1", 00:14:47.844 "aliases": [ 00:14:47.844 "649dd7ec-e339-410e-a0f9-6efbb12bb574" 00:14:47.844 ], 00:14:47.844 "product_name": "Malloc disk", 00:14:47.844 "block_size": 512, 00:14:47.844 "num_blocks": 65536, 00:14:47.844 "uuid": "649dd7ec-e339-410e-a0f9-6efbb12bb574", 00:14:47.844 "assigned_rate_limits": { 00:14:47.844 "rw_ios_per_sec": 0, 00:14:47.844 "rw_mbytes_per_sec": 0, 00:14:47.844 "r_mbytes_per_sec": 0, 00:14:47.844 "w_mbytes_per_sec": 0 00:14:47.844 }, 00:14:47.844 "claimed": true, 00:14:47.844 "claim_type": "exclusive_write", 00:14:47.844 "zoned": false, 00:14:47.844 "supported_io_types": { 00:14:47.844 "read": true, 00:14:47.844 "write": true, 00:14:47.844 "unmap": true, 00:14:47.844 "flush": true, 00:14:47.844 "reset": true, 00:14:47.844 "nvme_admin": false, 00:14:47.844 "nvme_io": false, 00:14:47.844 "nvme_io_md": false, 00:14:47.844 "write_zeroes": true, 00:14:47.844 "zcopy": true, 00:14:47.844 "get_zone_info": false, 00:14:47.844 "zone_management": false, 00:14:47.844 "zone_append": false, 00:14:47.844 "compare": false, 00:14:47.844 "compare_and_write": false, 00:14:47.844 "abort": true, 00:14:47.844 "seek_hole": false, 00:14:47.844 "seek_data": false, 00:14:47.844 "copy": true, 00:14:47.844 "nvme_iov_md": false 00:14:47.844 }, 00:14:47.844 "memory_domains": [ 00:14:47.844 { 00:14:47.844 "dma_device_id": "system", 00:14:47.844 "dma_device_type": 1 00:14:47.844 }, 00:14:47.844 { 00:14:47.844 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:47.844 "dma_device_type": 2 00:14:47.844 } 00:14:47.844 ], 00:14:47.844 "driver_specific": {} 00:14:47.844 } 00:14:47.844 ] 00:14:47.844 15:19:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.844 15:19:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:47.844 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:47.844 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:47.844 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:47.844 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:47.844 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:47.844 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:47.844 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:47.844 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:47.844 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:47.844 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:47.844 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:47.844 15:19:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.844 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:47.844 15:19:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:47.844 15:19:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.844 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:47.844 "name": "Existed_Raid", 00:14:47.844 "uuid": "900e6a7e-077f-474d-8a65-557a0dfacc8e", 00:14:47.844 "strip_size_kb": 64, 00:14:47.844 "state": "configuring", 00:14:47.844 "raid_level": "raid5f", 00:14:47.844 "superblock": true, 00:14:47.844 "num_base_bdevs": 4, 00:14:47.844 "num_base_bdevs_discovered": 3, 00:14:47.844 "num_base_bdevs_operational": 4, 00:14:47.844 "base_bdevs_list": [ 00:14:47.844 { 00:14:47.844 "name": "BaseBdev1", 00:14:47.844 "uuid": "649dd7ec-e339-410e-a0f9-6efbb12bb574", 00:14:47.844 "is_configured": true, 00:14:47.844 "data_offset": 2048, 00:14:47.844 "data_size": 63488 00:14:47.844 }, 00:14:47.844 { 00:14:47.844 "name": null, 00:14:47.844 "uuid": "c4357d34-75bf-4bdc-b44a-b72f3409c74a", 00:14:47.844 "is_configured": false, 00:14:47.844 "data_offset": 0, 00:14:47.844 "data_size": 63488 00:14:47.844 }, 00:14:47.844 { 00:14:47.844 "name": "BaseBdev3", 00:14:47.844 "uuid": "b232d9cd-feb4-49a2-969c-b9928c1339bf", 00:14:47.844 "is_configured": true, 00:14:47.844 "data_offset": 2048, 00:14:47.844 "data_size": 63488 00:14:47.844 }, 00:14:47.844 { 00:14:47.844 "name": "BaseBdev4", 00:14:47.844 "uuid": "ebaf47b1-0cc0-4ae9-88fb-23d780b39614", 00:14:47.844 "is_configured": true, 00:14:47.844 "data_offset": 2048, 00:14:47.844 "data_size": 63488 00:14:47.844 } 00:14:47.844 ] 00:14:47.844 }' 00:14:47.844 15:19:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:47.844 15:19:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:48.104 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:48.104 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:48.104 15:19:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.104 15:19:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:48.104 15:19:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.364 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:14:48.364 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:14:48.364 15:19:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.364 15:19:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:48.364 [2024-11-27 15:19:16.227540] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:48.364 15:19:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.364 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:48.364 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:48.364 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:48.364 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:48.364 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:48.364 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:48.364 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:48.364 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:48.364 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:48.364 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:48.364 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:48.364 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:48.364 15:19:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.364 15:19:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:48.364 15:19:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.364 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:48.364 "name": "Existed_Raid", 00:14:48.364 "uuid": "900e6a7e-077f-474d-8a65-557a0dfacc8e", 00:14:48.364 "strip_size_kb": 64, 00:14:48.364 "state": "configuring", 00:14:48.364 "raid_level": "raid5f", 00:14:48.364 "superblock": true, 00:14:48.364 "num_base_bdevs": 4, 00:14:48.364 "num_base_bdevs_discovered": 2, 00:14:48.364 "num_base_bdevs_operational": 4, 00:14:48.364 "base_bdevs_list": [ 00:14:48.364 { 00:14:48.364 "name": "BaseBdev1", 00:14:48.364 "uuid": "649dd7ec-e339-410e-a0f9-6efbb12bb574", 00:14:48.364 "is_configured": true, 00:14:48.364 "data_offset": 2048, 00:14:48.364 "data_size": 63488 00:14:48.364 }, 00:14:48.364 { 00:14:48.364 "name": null, 00:14:48.364 "uuid": "c4357d34-75bf-4bdc-b44a-b72f3409c74a", 00:14:48.364 "is_configured": false, 00:14:48.364 "data_offset": 0, 00:14:48.364 "data_size": 63488 00:14:48.364 }, 00:14:48.364 { 00:14:48.364 "name": null, 00:14:48.364 "uuid": "b232d9cd-feb4-49a2-969c-b9928c1339bf", 00:14:48.364 "is_configured": false, 00:14:48.364 "data_offset": 0, 00:14:48.364 "data_size": 63488 00:14:48.364 }, 00:14:48.364 { 00:14:48.364 "name": "BaseBdev4", 00:14:48.364 "uuid": "ebaf47b1-0cc0-4ae9-88fb-23d780b39614", 00:14:48.364 "is_configured": true, 00:14:48.364 "data_offset": 2048, 00:14:48.364 "data_size": 63488 00:14:48.364 } 00:14:48.364 ] 00:14:48.364 }' 00:14:48.364 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:48.364 15:19:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:48.624 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:48.624 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:48.624 15:19:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.624 15:19:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:48.624 15:19:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.624 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:14:48.624 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:14:48.624 15:19:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.624 15:19:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:48.624 [2024-11-27 15:19:16.699071] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:48.624 15:19:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.624 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:48.624 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:48.624 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:48.624 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:48.624 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:48.624 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:48.624 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:48.624 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:48.624 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:48.624 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:48.624 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:48.624 15:19:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.624 15:19:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:48.624 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:48.624 15:19:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.884 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:48.884 "name": "Existed_Raid", 00:14:48.884 "uuid": "900e6a7e-077f-474d-8a65-557a0dfacc8e", 00:14:48.885 "strip_size_kb": 64, 00:14:48.885 "state": "configuring", 00:14:48.885 "raid_level": "raid5f", 00:14:48.885 "superblock": true, 00:14:48.885 "num_base_bdevs": 4, 00:14:48.885 "num_base_bdevs_discovered": 3, 00:14:48.885 "num_base_bdevs_operational": 4, 00:14:48.885 "base_bdevs_list": [ 00:14:48.885 { 00:14:48.885 "name": "BaseBdev1", 00:14:48.885 "uuid": "649dd7ec-e339-410e-a0f9-6efbb12bb574", 00:14:48.885 "is_configured": true, 00:14:48.885 "data_offset": 2048, 00:14:48.885 "data_size": 63488 00:14:48.885 }, 00:14:48.885 { 00:14:48.885 "name": null, 00:14:48.885 "uuid": "c4357d34-75bf-4bdc-b44a-b72f3409c74a", 00:14:48.885 "is_configured": false, 00:14:48.885 "data_offset": 0, 00:14:48.885 "data_size": 63488 00:14:48.885 }, 00:14:48.885 { 00:14:48.885 "name": "BaseBdev3", 00:14:48.885 "uuid": "b232d9cd-feb4-49a2-969c-b9928c1339bf", 00:14:48.885 "is_configured": true, 00:14:48.885 "data_offset": 2048, 00:14:48.885 "data_size": 63488 00:14:48.885 }, 00:14:48.885 { 00:14:48.885 "name": "BaseBdev4", 00:14:48.885 "uuid": "ebaf47b1-0cc0-4ae9-88fb-23d780b39614", 00:14:48.885 "is_configured": true, 00:14:48.885 "data_offset": 2048, 00:14:48.885 "data_size": 63488 00:14:48.885 } 00:14:48.885 ] 00:14:48.885 }' 00:14:48.885 15:19:16 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:48.885 15:19:16 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:49.145 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:49.145 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:49.145 15:19:17 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.145 15:19:17 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:49.145 15:19:17 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.145 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:14:49.145 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:49.145 15:19:17 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.145 15:19:17 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:49.145 [2024-11-27 15:19:17.146277] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:49.145 15:19:17 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.145 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:49.145 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:49.145 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:49.145 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:49.145 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:49.145 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:49.145 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:49.145 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:49.145 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:49.145 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:49.145 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:49.145 15:19:17 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.145 15:19:17 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:49.145 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:49.145 15:19:17 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.145 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:49.145 "name": "Existed_Raid", 00:14:49.145 "uuid": "900e6a7e-077f-474d-8a65-557a0dfacc8e", 00:14:49.145 "strip_size_kb": 64, 00:14:49.145 "state": "configuring", 00:14:49.145 "raid_level": "raid5f", 00:14:49.145 "superblock": true, 00:14:49.145 "num_base_bdevs": 4, 00:14:49.145 "num_base_bdevs_discovered": 2, 00:14:49.145 "num_base_bdevs_operational": 4, 00:14:49.145 "base_bdevs_list": [ 00:14:49.145 { 00:14:49.145 "name": null, 00:14:49.145 "uuid": "649dd7ec-e339-410e-a0f9-6efbb12bb574", 00:14:49.145 "is_configured": false, 00:14:49.145 "data_offset": 0, 00:14:49.145 "data_size": 63488 00:14:49.145 }, 00:14:49.145 { 00:14:49.145 "name": null, 00:14:49.145 "uuid": "c4357d34-75bf-4bdc-b44a-b72f3409c74a", 00:14:49.145 "is_configured": false, 00:14:49.145 "data_offset": 0, 00:14:49.145 "data_size": 63488 00:14:49.145 }, 00:14:49.145 { 00:14:49.145 "name": "BaseBdev3", 00:14:49.145 "uuid": "b232d9cd-feb4-49a2-969c-b9928c1339bf", 00:14:49.145 "is_configured": true, 00:14:49.145 "data_offset": 2048, 00:14:49.145 "data_size": 63488 00:14:49.145 }, 00:14:49.145 { 00:14:49.145 "name": "BaseBdev4", 00:14:49.145 "uuid": "ebaf47b1-0cc0-4ae9-88fb-23d780b39614", 00:14:49.145 "is_configured": true, 00:14:49.145 "data_offset": 2048, 00:14:49.145 "data_size": 63488 00:14:49.145 } 00:14:49.145 ] 00:14:49.145 }' 00:14:49.145 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:49.145 15:19:17 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:49.715 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:49.715 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:49.715 15:19:17 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.715 15:19:17 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:49.715 15:19:17 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.715 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:14:49.715 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:14:49.715 15:19:17 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.715 15:19:17 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:49.715 [2024-11-27 15:19:17.657781] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:49.715 15:19:17 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.715 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:49.715 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:49.715 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:49.715 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:49.715 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:49.715 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:49.715 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:49.715 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:49.715 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:49.715 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:49.715 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:49.715 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:49.715 15:19:17 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.715 15:19:17 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:49.715 15:19:17 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.715 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:49.715 "name": "Existed_Raid", 00:14:49.715 "uuid": "900e6a7e-077f-474d-8a65-557a0dfacc8e", 00:14:49.715 "strip_size_kb": 64, 00:14:49.715 "state": "configuring", 00:14:49.715 "raid_level": "raid5f", 00:14:49.715 "superblock": true, 00:14:49.715 "num_base_bdevs": 4, 00:14:49.715 "num_base_bdevs_discovered": 3, 00:14:49.715 "num_base_bdevs_operational": 4, 00:14:49.715 "base_bdevs_list": [ 00:14:49.715 { 00:14:49.715 "name": null, 00:14:49.715 "uuid": "649dd7ec-e339-410e-a0f9-6efbb12bb574", 00:14:49.715 "is_configured": false, 00:14:49.715 "data_offset": 0, 00:14:49.715 "data_size": 63488 00:14:49.715 }, 00:14:49.715 { 00:14:49.715 "name": "BaseBdev2", 00:14:49.715 "uuid": "c4357d34-75bf-4bdc-b44a-b72f3409c74a", 00:14:49.715 "is_configured": true, 00:14:49.716 "data_offset": 2048, 00:14:49.716 "data_size": 63488 00:14:49.716 }, 00:14:49.716 { 00:14:49.716 "name": "BaseBdev3", 00:14:49.716 "uuid": "b232d9cd-feb4-49a2-969c-b9928c1339bf", 00:14:49.716 "is_configured": true, 00:14:49.716 "data_offset": 2048, 00:14:49.716 "data_size": 63488 00:14:49.716 }, 00:14:49.716 { 00:14:49.716 "name": "BaseBdev4", 00:14:49.716 "uuid": "ebaf47b1-0cc0-4ae9-88fb-23d780b39614", 00:14:49.716 "is_configured": true, 00:14:49.716 "data_offset": 2048, 00:14:49.716 "data_size": 63488 00:14:49.716 } 00:14:49.716 ] 00:14:49.716 }' 00:14:49.716 15:19:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:49.716 15:19:17 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:50.285 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:50.285 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:50.285 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.285 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:50.285 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.285 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:14:50.285 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:50.285 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 649dd7ec-e339-410e-a0f9-6efbb12bb574 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:50.286 [2024-11-27 15:19:18.180848] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:14:50.286 [2024-11-27 15:19:18.181179] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:14:50.286 [2024-11-27 15:19:18.181240] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:50.286 [2024-11-27 15:19:18.181593] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:14:50.286 NewBaseBdev 00:14:50.286 [2024-11-27 15:19:18.182136] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:14:50.286 [2024-11-27 15:19:18.182162] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006d00 00:14:50.286 [2024-11-27 15:19:18.182291] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:50.286 [ 00:14:50.286 { 00:14:50.286 "name": "NewBaseBdev", 00:14:50.286 "aliases": [ 00:14:50.286 "649dd7ec-e339-410e-a0f9-6efbb12bb574" 00:14:50.286 ], 00:14:50.286 "product_name": "Malloc disk", 00:14:50.286 "block_size": 512, 00:14:50.286 "num_blocks": 65536, 00:14:50.286 "uuid": "649dd7ec-e339-410e-a0f9-6efbb12bb574", 00:14:50.286 "assigned_rate_limits": { 00:14:50.286 "rw_ios_per_sec": 0, 00:14:50.286 "rw_mbytes_per_sec": 0, 00:14:50.286 "r_mbytes_per_sec": 0, 00:14:50.286 "w_mbytes_per_sec": 0 00:14:50.286 }, 00:14:50.286 "claimed": true, 00:14:50.286 "claim_type": "exclusive_write", 00:14:50.286 "zoned": false, 00:14:50.286 "supported_io_types": { 00:14:50.286 "read": true, 00:14:50.286 "write": true, 00:14:50.286 "unmap": true, 00:14:50.286 "flush": true, 00:14:50.286 "reset": true, 00:14:50.286 "nvme_admin": false, 00:14:50.286 "nvme_io": false, 00:14:50.286 "nvme_io_md": false, 00:14:50.286 "write_zeroes": true, 00:14:50.286 "zcopy": true, 00:14:50.286 "get_zone_info": false, 00:14:50.286 "zone_management": false, 00:14:50.286 "zone_append": false, 00:14:50.286 "compare": false, 00:14:50.286 "compare_and_write": false, 00:14:50.286 "abort": true, 00:14:50.286 "seek_hole": false, 00:14:50.286 "seek_data": false, 00:14:50.286 "copy": true, 00:14:50.286 "nvme_iov_md": false 00:14:50.286 }, 00:14:50.286 "memory_domains": [ 00:14:50.286 { 00:14:50.286 "dma_device_id": "system", 00:14:50.286 "dma_device_type": 1 00:14:50.286 }, 00:14:50.286 { 00:14:50.286 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:50.286 "dma_device_type": 2 00:14:50.286 } 00:14:50.286 ], 00:14:50.286 "driver_specific": {} 00:14:50.286 } 00:14:50.286 ] 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:50.286 "name": "Existed_Raid", 00:14:50.286 "uuid": "900e6a7e-077f-474d-8a65-557a0dfacc8e", 00:14:50.286 "strip_size_kb": 64, 00:14:50.286 "state": "online", 00:14:50.286 "raid_level": "raid5f", 00:14:50.286 "superblock": true, 00:14:50.286 "num_base_bdevs": 4, 00:14:50.286 "num_base_bdevs_discovered": 4, 00:14:50.286 "num_base_bdevs_operational": 4, 00:14:50.286 "base_bdevs_list": [ 00:14:50.286 { 00:14:50.286 "name": "NewBaseBdev", 00:14:50.286 "uuid": "649dd7ec-e339-410e-a0f9-6efbb12bb574", 00:14:50.286 "is_configured": true, 00:14:50.286 "data_offset": 2048, 00:14:50.286 "data_size": 63488 00:14:50.286 }, 00:14:50.286 { 00:14:50.286 "name": "BaseBdev2", 00:14:50.286 "uuid": "c4357d34-75bf-4bdc-b44a-b72f3409c74a", 00:14:50.286 "is_configured": true, 00:14:50.286 "data_offset": 2048, 00:14:50.286 "data_size": 63488 00:14:50.286 }, 00:14:50.286 { 00:14:50.286 "name": "BaseBdev3", 00:14:50.286 "uuid": "b232d9cd-feb4-49a2-969c-b9928c1339bf", 00:14:50.286 "is_configured": true, 00:14:50.286 "data_offset": 2048, 00:14:50.286 "data_size": 63488 00:14:50.286 }, 00:14:50.286 { 00:14:50.286 "name": "BaseBdev4", 00:14:50.286 "uuid": "ebaf47b1-0cc0-4ae9-88fb-23d780b39614", 00:14:50.286 "is_configured": true, 00:14:50.286 "data_offset": 2048, 00:14:50.286 "data_size": 63488 00:14:50.286 } 00:14:50.286 ] 00:14:50.286 }' 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:50.286 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:50.856 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:14:50.856 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:50.856 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:50.856 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:50.856 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:14:50.856 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:50.856 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:50.856 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:50.856 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.856 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:50.856 [2024-11-27 15:19:18.696296] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:50.856 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.856 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:50.856 "name": "Existed_Raid", 00:14:50.856 "aliases": [ 00:14:50.856 "900e6a7e-077f-474d-8a65-557a0dfacc8e" 00:14:50.856 ], 00:14:50.856 "product_name": "Raid Volume", 00:14:50.856 "block_size": 512, 00:14:50.856 "num_blocks": 190464, 00:14:50.856 "uuid": "900e6a7e-077f-474d-8a65-557a0dfacc8e", 00:14:50.856 "assigned_rate_limits": { 00:14:50.856 "rw_ios_per_sec": 0, 00:14:50.856 "rw_mbytes_per_sec": 0, 00:14:50.856 "r_mbytes_per_sec": 0, 00:14:50.856 "w_mbytes_per_sec": 0 00:14:50.856 }, 00:14:50.856 "claimed": false, 00:14:50.856 "zoned": false, 00:14:50.856 "supported_io_types": { 00:14:50.856 "read": true, 00:14:50.856 "write": true, 00:14:50.856 "unmap": false, 00:14:50.857 "flush": false, 00:14:50.857 "reset": true, 00:14:50.857 "nvme_admin": false, 00:14:50.857 "nvme_io": false, 00:14:50.857 "nvme_io_md": false, 00:14:50.857 "write_zeroes": true, 00:14:50.857 "zcopy": false, 00:14:50.857 "get_zone_info": false, 00:14:50.857 "zone_management": false, 00:14:50.857 "zone_append": false, 00:14:50.857 "compare": false, 00:14:50.857 "compare_and_write": false, 00:14:50.857 "abort": false, 00:14:50.857 "seek_hole": false, 00:14:50.857 "seek_data": false, 00:14:50.857 "copy": false, 00:14:50.857 "nvme_iov_md": false 00:14:50.857 }, 00:14:50.857 "driver_specific": { 00:14:50.857 "raid": { 00:14:50.857 "uuid": "900e6a7e-077f-474d-8a65-557a0dfacc8e", 00:14:50.857 "strip_size_kb": 64, 00:14:50.857 "state": "online", 00:14:50.857 "raid_level": "raid5f", 00:14:50.857 "superblock": true, 00:14:50.857 "num_base_bdevs": 4, 00:14:50.857 "num_base_bdevs_discovered": 4, 00:14:50.857 "num_base_bdevs_operational": 4, 00:14:50.857 "base_bdevs_list": [ 00:14:50.857 { 00:14:50.857 "name": "NewBaseBdev", 00:14:50.857 "uuid": "649dd7ec-e339-410e-a0f9-6efbb12bb574", 00:14:50.857 "is_configured": true, 00:14:50.857 "data_offset": 2048, 00:14:50.857 "data_size": 63488 00:14:50.857 }, 00:14:50.857 { 00:14:50.857 "name": "BaseBdev2", 00:14:50.857 "uuid": "c4357d34-75bf-4bdc-b44a-b72f3409c74a", 00:14:50.857 "is_configured": true, 00:14:50.857 "data_offset": 2048, 00:14:50.857 "data_size": 63488 00:14:50.857 }, 00:14:50.857 { 00:14:50.857 "name": "BaseBdev3", 00:14:50.857 "uuid": "b232d9cd-feb4-49a2-969c-b9928c1339bf", 00:14:50.857 "is_configured": true, 00:14:50.857 "data_offset": 2048, 00:14:50.857 "data_size": 63488 00:14:50.857 }, 00:14:50.857 { 00:14:50.857 "name": "BaseBdev4", 00:14:50.857 "uuid": "ebaf47b1-0cc0-4ae9-88fb-23d780b39614", 00:14:50.857 "is_configured": true, 00:14:50.857 "data_offset": 2048, 00:14:50.857 "data_size": 63488 00:14:50.857 } 00:14:50.857 ] 00:14:50.857 } 00:14:50.857 } 00:14:50.857 }' 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:14:50.857 BaseBdev2 00:14:50.857 BaseBdev3 00:14:50.857 BaseBdev4' 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:50.857 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:51.118 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:51.118 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:51.118 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:51.118 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:51.118 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:51.118 [2024-11-27 15:19:18.995579] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:51.118 [2024-11-27 15:19:18.995657] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:51.118 [2024-11-27 15:19:18.995758] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:51.118 [2024-11-27 15:19:18.996094] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:51.118 [2024-11-27 15:19:18.996160] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name Existed_Raid, state offline 00:14:51.118 15:19:18 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:51.118 15:19:18 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 94085 00:14:51.118 15:19:19 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 94085 ']' 00:14:51.118 15:19:19 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 94085 00:14:51.118 15:19:19 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:14:51.118 15:19:19 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:51.118 15:19:19 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 94085 00:14:51.118 15:19:19 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:14:51.118 killing process with pid 94085 00:14:51.118 15:19:19 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:14:51.118 15:19:19 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 94085' 00:14:51.118 15:19:19 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 94085 00:14:51.118 [2024-11-27 15:19:19.045300] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:51.118 15:19:19 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 94085 00:14:51.118 [2024-11-27 15:19:19.126404] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:51.378 15:19:19 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:14:51.378 00:14:51.378 real 0m9.783s 00:14:51.378 user 0m16.406s 00:14:51.378 sys 0m2.172s 00:14:51.378 15:19:19 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:51.378 15:19:19 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:51.378 ************************************ 00:14:51.378 END TEST raid5f_state_function_test_sb 00:14:51.378 ************************************ 00:14:51.638 15:19:19 bdev_raid -- bdev/bdev_raid.sh@988 -- # run_test raid5f_superblock_test raid_superblock_test raid5f 4 00:14:51.638 15:19:19 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:14:51.638 15:19:19 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:51.638 15:19:19 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:51.638 ************************************ 00:14:51.638 START TEST raid5f_superblock_test 00:14:51.638 ************************************ 00:14:51.638 15:19:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid5f 4 00:14:51.638 15:19:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid5f 00:14:51.638 15:19:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:14:51.638 15:19:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:14:51.638 15:19:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:14:51.638 15:19:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:14:51.638 15:19:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:14:51.638 15:19:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:14:51.638 15:19:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:14:51.638 15:19:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:14:51.638 15:19:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:14:51.638 15:19:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:14:51.638 15:19:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:14:51.638 15:19:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:14:51.638 15:19:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid5f '!=' raid1 ']' 00:14:51.638 15:19:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:14:51.638 15:19:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:14:51.638 15:19:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=94736 00:14:51.638 15:19:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:14:51.638 15:19:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 94736 00:14:51.638 15:19:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 94736 ']' 00:14:51.638 15:19:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:51.638 15:19:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:51.638 15:19:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:51.638 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:51.638 15:19:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:51.638 15:19:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:51.638 [2024-11-27 15:19:19.633790] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:14:51.638 [2024-11-27 15:19:19.634020] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid94736 ] 00:14:51.897 [2024-11-27 15:19:19.802668] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:51.897 [2024-11-27 15:19:19.841006] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:51.897 [2024-11-27 15:19:19.917998] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:51.897 [2024-11-27 15:19:19.918143] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:52.468 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:52.468 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:14:52.468 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:14:52.468 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:52.468 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:14:52.468 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:14:52.468 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:14:52.468 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:52.468 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:14:52.468 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:52.468 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:14:52.468 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:52.468 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.468 malloc1 00:14:52.468 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:52.468 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:14:52.468 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:52.468 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.468 [2024-11-27 15:19:20.499856] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:14:52.468 [2024-11-27 15:19:20.500067] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:52.468 [2024-11-27 15:19:20.500116] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:14:52.468 [2024-11-27 15:19:20.500180] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:52.468 [2024-11-27 15:19:20.502666] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:52.468 [2024-11-27 15:19:20.502756] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:14:52.468 pt1 00:14:52.468 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:52.468 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:14:52.468 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:52.468 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:14:52.468 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:14:52.468 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:14:52.468 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:52.468 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:14:52.468 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:52.469 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:14:52.469 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:52.469 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.469 malloc2 00:14:52.469 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:52.469 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:52.469 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:52.469 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.469 [2024-11-27 15:19:20.538865] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:52.469 [2024-11-27 15:19:20.539001] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:52.469 [2024-11-27 15:19:20.539044] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:14:52.469 [2024-11-27 15:19:20.539086] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:52.469 [2024-11-27 15:19:20.541619] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:52.469 [2024-11-27 15:19:20.541709] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:52.469 pt2 00:14:52.469 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:52.469 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:14:52.469 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:52.469 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:14:52.469 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:14:52.469 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:14:52.469 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:52.469 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:14:52.469 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:52.469 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:14:52.469 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:52.469 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.469 malloc3 00:14:52.469 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:52.469 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:14:52.469 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:52.469 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.730 [2024-11-27 15:19:20.577561] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:14:52.730 [2024-11-27 15:19:20.577675] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:52.730 [2024-11-27 15:19:20.577715] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:14:52.730 [2024-11-27 15:19:20.577751] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:52.730 [2024-11-27 15:19:20.580063] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:52.730 [2024-11-27 15:19:20.580145] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:14:52.730 pt3 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.730 malloc4 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.730 [2024-11-27 15:19:20.620050] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:14:52.730 [2024-11-27 15:19:20.620175] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:52.730 [2024-11-27 15:19:20.620217] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:14:52.730 [2024-11-27 15:19:20.620261] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:52.730 [2024-11-27 15:19:20.622565] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:52.730 [2024-11-27 15:19:20.622647] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:14:52.730 pt4 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.730 [2024-11-27 15:19:20.632128] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:14:52.730 [2024-11-27 15:19:20.634227] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:52.730 [2024-11-27 15:19:20.634360] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:52.730 [2024-11-27 15:19:20.634437] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:14:52.730 [2024-11-27 15:19:20.634655] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:14:52.730 [2024-11-27 15:19:20.634720] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:52.730 [2024-11-27 15:19:20.635064] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:14:52.730 [2024-11-27 15:19:20.635652] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:14:52.730 [2024-11-27 15:19:20.635707] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:14:52.730 [2024-11-27 15:19:20.635880] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:52.730 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:52.731 "name": "raid_bdev1", 00:14:52.731 "uuid": "d7073afc-9d76-4702-a72e-3a57959cdd6b", 00:14:52.731 "strip_size_kb": 64, 00:14:52.731 "state": "online", 00:14:52.731 "raid_level": "raid5f", 00:14:52.731 "superblock": true, 00:14:52.731 "num_base_bdevs": 4, 00:14:52.731 "num_base_bdevs_discovered": 4, 00:14:52.731 "num_base_bdevs_operational": 4, 00:14:52.731 "base_bdevs_list": [ 00:14:52.731 { 00:14:52.731 "name": "pt1", 00:14:52.731 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:52.731 "is_configured": true, 00:14:52.731 "data_offset": 2048, 00:14:52.731 "data_size": 63488 00:14:52.731 }, 00:14:52.731 { 00:14:52.731 "name": "pt2", 00:14:52.731 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:52.731 "is_configured": true, 00:14:52.731 "data_offset": 2048, 00:14:52.731 "data_size": 63488 00:14:52.731 }, 00:14:52.731 { 00:14:52.731 "name": "pt3", 00:14:52.731 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:52.731 "is_configured": true, 00:14:52.731 "data_offset": 2048, 00:14:52.731 "data_size": 63488 00:14:52.731 }, 00:14:52.731 { 00:14:52.731 "name": "pt4", 00:14:52.731 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:52.731 "is_configured": true, 00:14:52.731 "data_offset": 2048, 00:14:52.731 "data_size": 63488 00:14:52.731 } 00:14:52.731 ] 00:14:52.731 }' 00:14:52.731 15:19:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:52.731 15:19:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.991 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:14:52.991 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:14:52.991 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:52.991 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:52.991 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:52.991 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:52.991 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:52.991 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:52.991 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:52.991 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.991 [2024-11-27 15:19:21.075716] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:52.991 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:53.251 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:53.251 "name": "raid_bdev1", 00:14:53.251 "aliases": [ 00:14:53.251 "d7073afc-9d76-4702-a72e-3a57959cdd6b" 00:14:53.251 ], 00:14:53.251 "product_name": "Raid Volume", 00:14:53.251 "block_size": 512, 00:14:53.251 "num_blocks": 190464, 00:14:53.251 "uuid": "d7073afc-9d76-4702-a72e-3a57959cdd6b", 00:14:53.251 "assigned_rate_limits": { 00:14:53.251 "rw_ios_per_sec": 0, 00:14:53.251 "rw_mbytes_per_sec": 0, 00:14:53.251 "r_mbytes_per_sec": 0, 00:14:53.251 "w_mbytes_per_sec": 0 00:14:53.251 }, 00:14:53.251 "claimed": false, 00:14:53.251 "zoned": false, 00:14:53.251 "supported_io_types": { 00:14:53.251 "read": true, 00:14:53.251 "write": true, 00:14:53.251 "unmap": false, 00:14:53.251 "flush": false, 00:14:53.251 "reset": true, 00:14:53.251 "nvme_admin": false, 00:14:53.251 "nvme_io": false, 00:14:53.251 "nvme_io_md": false, 00:14:53.251 "write_zeroes": true, 00:14:53.251 "zcopy": false, 00:14:53.251 "get_zone_info": false, 00:14:53.251 "zone_management": false, 00:14:53.251 "zone_append": false, 00:14:53.251 "compare": false, 00:14:53.251 "compare_and_write": false, 00:14:53.251 "abort": false, 00:14:53.251 "seek_hole": false, 00:14:53.251 "seek_data": false, 00:14:53.252 "copy": false, 00:14:53.252 "nvme_iov_md": false 00:14:53.252 }, 00:14:53.252 "driver_specific": { 00:14:53.252 "raid": { 00:14:53.252 "uuid": "d7073afc-9d76-4702-a72e-3a57959cdd6b", 00:14:53.252 "strip_size_kb": 64, 00:14:53.252 "state": "online", 00:14:53.252 "raid_level": "raid5f", 00:14:53.252 "superblock": true, 00:14:53.252 "num_base_bdevs": 4, 00:14:53.252 "num_base_bdevs_discovered": 4, 00:14:53.252 "num_base_bdevs_operational": 4, 00:14:53.252 "base_bdevs_list": [ 00:14:53.252 { 00:14:53.252 "name": "pt1", 00:14:53.252 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:53.252 "is_configured": true, 00:14:53.252 "data_offset": 2048, 00:14:53.252 "data_size": 63488 00:14:53.252 }, 00:14:53.252 { 00:14:53.252 "name": "pt2", 00:14:53.252 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:53.252 "is_configured": true, 00:14:53.252 "data_offset": 2048, 00:14:53.252 "data_size": 63488 00:14:53.252 }, 00:14:53.252 { 00:14:53.252 "name": "pt3", 00:14:53.252 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:53.252 "is_configured": true, 00:14:53.252 "data_offset": 2048, 00:14:53.252 "data_size": 63488 00:14:53.252 }, 00:14:53.252 { 00:14:53.252 "name": "pt4", 00:14:53.252 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:53.252 "is_configured": true, 00:14:53.252 "data_offset": 2048, 00:14:53.252 "data_size": 63488 00:14:53.252 } 00:14:53.252 ] 00:14:53.252 } 00:14:53.252 } 00:14:53.252 }' 00:14:53.252 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:53.252 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:14:53.252 pt2 00:14:53.252 pt3 00:14:53.252 pt4' 00:14:53.252 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:53.252 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:53.252 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:53.252 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:14:53.252 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:53.252 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:53.252 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.252 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:53.252 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:53.252 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:53.252 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:53.252 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:14:53.252 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:53.252 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.252 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:53.252 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:53.252 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:53.252 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:53.252 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:53.252 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:53.252 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:14:53.252 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:53.252 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.252 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.512 [2024-11-27 15:19:21.427035] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=d7073afc-9d76-4702-a72e-3a57959cdd6b 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z d7073afc-9d76-4702-a72e-3a57959cdd6b ']' 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.512 [2024-11-27 15:19:21.474766] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:53.512 [2024-11-27 15:19:21.474845] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:53.512 [2024-11-27 15:19:21.474970] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:53.512 [2024-11-27 15:19:21.475083] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:53.512 [2024-11-27 15:19:21.475139] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:53.512 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:14:53.513 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:14:53.513 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:53.513 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.513 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:53.513 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:14:53.513 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:14:53.513 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:53.513 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.513 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:53.513 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:14:53.513 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:14:53.513 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:53.513 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.773 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:53.773 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:14:53.773 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:14:53.773 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:14:53.773 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:14:53.773 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:14:53.773 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:53.773 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:14:53.773 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:53.773 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:14:53.773 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:53.773 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.773 [2024-11-27 15:19:21.638518] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:14:53.773 [2024-11-27 15:19:21.640758] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:14:53.773 [2024-11-27 15:19:21.640859] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:14:53.773 [2024-11-27 15:19:21.640937] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:14:53.773 [2024-11-27 15:19:21.641026] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:14:53.773 [2024-11-27 15:19:21.641128] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:14:53.773 [2024-11-27 15:19:21.641242] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:14:53.773 [2024-11-27 15:19:21.641309] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:14:53.773 [2024-11-27 15:19:21.641367] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:53.773 [2024-11-27 15:19:21.641418] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state configuring 00:14:53.773 request: 00:14:53.773 { 00:14:53.773 "name": "raid_bdev1", 00:14:53.773 "raid_level": "raid5f", 00:14:53.773 "base_bdevs": [ 00:14:53.773 "malloc1", 00:14:53.773 "malloc2", 00:14:53.773 "malloc3", 00:14:53.773 "malloc4" 00:14:53.773 ], 00:14:53.773 "strip_size_kb": 64, 00:14:53.773 "superblock": false, 00:14:53.773 "method": "bdev_raid_create", 00:14:53.773 "req_id": 1 00:14:53.773 } 00:14:53.773 Got JSON-RPC error response 00:14:53.773 response: 00:14:53.773 { 00:14:53.773 "code": -17, 00:14:53.773 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:14:53.773 } 00:14:53.773 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:14:53.773 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:14:53.773 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:14:53.773 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:14:53.773 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:14:53.773 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:14:53.773 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:53.774 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:53.774 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.774 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:53.774 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:14:53.774 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:14:53.774 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:14:53.774 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:53.774 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.774 [2024-11-27 15:19:21.694364] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:14:53.774 [2024-11-27 15:19:21.694456] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:53.774 [2024-11-27 15:19:21.694498] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:14:53.774 [2024-11-27 15:19:21.694530] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:53.774 [2024-11-27 15:19:21.697003] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:53.774 [2024-11-27 15:19:21.697083] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:14:53.774 [2024-11-27 15:19:21.697180] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:14:53.774 [2024-11-27 15:19:21.697251] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:14:53.774 pt1 00:14:53.774 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:53.774 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 4 00:14:53.774 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:53.774 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:53.774 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:53.774 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:53.774 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:53.774 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:53.774 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:53.774 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:53.774 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:53.774 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:53.774 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:53.774 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:53.774 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.774 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:53.774 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:53.774 "name": "raid_bdev1", 00:14:53.774 "uuid": "d7073afc-9d76-4702-a72e-3a57959cdd6b", 00:14:53.774 "strip_size_kb": 64, 00:14:53.774 "state": "configuring", 00:14:53.774 "raid_level": "raid5f", 00:14:53.774 "superblock": true, 00:14:53.774 "num_base_bdevs": 4, 00:14:53.774 "num_base_bdevs_discovered": 1, 00:14:53.774 "num_base_bdevs_operational": 4, 00:14:53.774 "base_bdevs_list": [ 00:14:53.774 { 00:14:53.774 "name": "pt1", 00:14:53.774 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:53.774 "is_configured": true, 00:14:53.774 "data_offset": 2048, 00:14:53.774 "data_size": 63488 00:14:53.774 }, 00:14:53.774 { 00:14:53.774 "name": null, 00:14:53.774 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:53.774 "is_configured": false, 00:14:53.774 "data_offset": 2048, 00:14:53.774 "data_size": 63488 00:14:53.774 }, 00:14:53.774 { 00:14:53.774 "name": null, 00:14:53.774 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:53.774 "is_configured": false, 00:14:53.774 "data_offset": 2048, 00:14:53.774 "data_size": 63488 00:14:53.774 }, 00:14:53.774 { 00:14:53.774 "name": null, 00:14:53.774 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:53.774 "is_configured": false, 00:14:53.774 "data_offset": 2048, 00:14:53.774 "data_size": 63488 00:14:53.774 } 00:14:53.774 ] 00:14:53.774 }' 00:14:53.774 15:19:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:53.774 15:19:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.344 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:14:54.344 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:54.344 15:19:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.344 15:19:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.344 [2024-11-27 15:19:22.145684] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:54.344 [2024-11-27 15:19:22.145797] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:54.344 [2024-11-27 15:19:22.145840] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:14:54.344 [2024-11-27 15:19:22.145873] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:54.344 [2024-11-27 15:19:22.146328] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:54.344 [2024-11-27 15:19:22.146394] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:54.344 [2024-11-27 15:19:22.146496] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:14:54.344 [2024-11-27 15:19:22.146555] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:54.344 pt2 00:14:54.344 15:19:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.344 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:14:54.344 15:19:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.344 15:19:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.344 [2024-11-27 15:19:22.157675] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:14:54.344 15:19:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.344 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 4 00:14:54.345 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:54.345 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:54.345 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:54.345 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:54.345 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:54.345 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:54.345 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:54.345 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:54.345 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:54.345 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:54.345 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:54.345 15:19:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.345 15:19:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.345 15:19:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.345 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:54.345 "name": "raid_bdev1", 00:14:54.345 "uuid": "d7073afc-9d76-4702-a72e-3a57959cdd6b", 00:14:54.345 "strip_size_kb": 64, 00:14:54.345 "state": "configuring", 00:14:54.345 "raid_level": "raid5f", 00:14:54.345 "superblock": true, 00:14:54.345 "num_base_bdevs": 4, 00:14:54.345 "num_base_bdevs_discovered": 1, 00:14:54.345 "num_base_bdevs_operational": 4, 00:14:54.345 "base_bdevs_list": [ 00:14:54.345 { 00:14:54.345 "name": "pt1", 00:14:54.345 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:54.345 "is_configured": true, 00:14:54.345 "data_offset": 2048, 00:14:54.345 "data_size": 63488 00:14:54.345 }, 00:14:54.345 { 00:14:54.345 "name": null, 00:14:54.345 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:54.345 "is_configured": false, 00:14:54.345 "data_offset": 0, 00:14:54.345 "data_size": 63488 00:14:54.345 }, 00:14:54.345 { 00:14:54.345 "name": null, 00:14:54.345 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:54.345 "is_configured": false, 00:14:54.345 "data_offset": 2048, 00:14:54.345 "data_size": 63488 00:14:54.345 }, 00:14:54.345 { 00:14:54.345 "name": null, 00:14:54.345 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:54.345 "is_configured": false, 00:14:54.345 "data_offset": 2048, 00:14:54.345 "data_size": 63488 00:14:54.345 } 00:14:54.345 ] 00:14:54.345 }' 00:14:54.345 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:54.345 15:19:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.610 [2024-11-27 15:19:22.601036] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:54.610 [2024-11-27 15:19:22.601153] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:54.610 [2024-11-27 15:19:22.601190] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:14:54.610 [2024-11-27 15:19:22.601236] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:54.610 [2024-11-27 15:19:22.601648] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:54.610 [2024-11-27 15:19:22.601717] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:54.610 [2024-11-27 15:19:22.601809] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:14:54.610 [2024-11-27 15:19:22.601864] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:54.610 pt2 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.610 [2024-11-27 15:19:22.613031] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:14:54.610 [2024-11-27 15:19:22.613130] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:54.610 [2024-11-27 15:19:22.613164] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:14:54.610 [2024-11-27 15:19:22.613209] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:54.610 [2024-11-27 15:19:22.613549] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:54.610 [2024-11-27 15:19:22.613615] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:14:54.610 [2024-11-27 15:19:22.613696] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:14:54.610 [2024-11-27 15:19:22.613760] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:54.610 pt3 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.610 [2024-11-27 15:19:22.625011] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:14:54.610 [2024-11-27 15:19:22.625110] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:54.610 [2024-11-27 15:19:22.625152] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:14:54.610 [2024-11-27 15:19:22.625185] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:54.610 [2024-11-27 15:19:22.625531] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:54.610 [2024-11-27 15:19:22.625597] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:14:54.610 [2024-11-27 15:19:22.625680] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:14:54.610 [2024-11-27 15:19:22.625734] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:14:54.610 [2024-11-27 15:19:22.625867] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:14:54.610 [2024-11-27 15:19:22.625930] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:54.610 [2024-11-27 15:19:22.626208] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:14:54.610 [2024-11-27 15:19:22.626768] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:14:54.610 [2024-11-27 15:19:22.626820] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:14:54.610 [2024-11-27 15:19:22.626979] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:54.610 pt4 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.610 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:54.610 "name": "raid_bdev1", 00:14:54.610 "uuid": "d7073afc-9d76-4702-a72e-3a57959cdd6b", 00:14:54.610 "strip_size_kb": 64, 00:14:54.610 "state": "online", 00:14:54.610 "raid_level": "raid5f", 00:14:54.610 "superblock": true, 00:14:54.610 "num_base_bdevs": 4, 00:14:54.610 "num_base_bdevs_discovered": 4, 00:14:54.610 "num_base_bdevs_operational": 4, 00:14:54.610 "base_bdevs_list": [ 00:14:54.610 { 00:14:54.610 "name": "pt1", 00:14:54.610 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:54.610 "is_configured": true, 00:14:54.610 "data_offset": 2048, 00:14:54.610 "data_size": 63488 00:14:54.610 }, 00:14:54.610 { 00:14:54.610 "name": "pt2", 00:14:54.610 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:54.611 "is_configured": true, 00:14:54.611 "data_offset": 2048, 00:14:54.611 "data_size": 63488 00:14:54.611 }, 00:14:54.611 { 00:14:54.611 "name": "pt3", 00:14:54.611 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:54.611 "is_configured": true, 00:14:54.611 "data_offset": 2048, 00:14:54.611 "data_size": 63488 00:14:54.611 }, 00:14:54.611 { 00:14:54.611 "name": "pt4", 00:14:54.611 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:54.611 "is_configured": true, 00:14:54.611 "data_offset": 2048, 00:14:54.611 "data_size": 63488 00:14:54.611 } 00:14:54.611 ] 00:14:54.611 }' 00:14:54.611 15:19:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:54.611 15:19:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:55.191 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:14:55.191 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:14:55.191 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:55.191 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:55.191 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:55.191 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:55.191 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:55.191 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:55.191 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:55.191 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:55.191 [2024-11-27 15:19:23.073178] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:55.191 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:55.191 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:55.191 "name": "raid_bdev1", 00:14:55.191 "aliases": [ 00:14:55.191 "d7073afc-9d76-4702-a72e-3a57959cdd6b" 00:14:55.191 ], 00:14:55.191 "product_name": "Raid Volume", 00:14:55.191 "block_size": 512, 00:14:55.191 "num_blocks": 190464, 00:14:55.191 "uuid": "d7073afc-9d76-4702-a72e-3a57959cdd6b", 00:14:55.191 "assigned_rate_limits": { 00:14:55.191 "rw_ios_per_sec": 0, 00:14:55.191 "rw_mbytes_per_sec": 0, 00:14:55.191 "r_mbytes_per_sec": 0, 00:14:55.191 "w_mbytes_per_sec": 0 00:14:55.191 }, 00:14:55.191 "claimed": false, 00:14:55.191 "zoned": false, 00:14:55.191 "supported_io_types": { 00:14:55.191 "read": true, 00:14:55.191 "write": true, 00:14:55.191 "unmap": false, 00:14:55.191 "flush": false, 00:14:55.191 "reset": true, 00:14:55.191 "nvme_admin": false, 00:14:55.191 "nvme_io": false, 00:14:55.191 "nvme_io_md": false, 00:14:55.191 "write_zeroes": true, 00:14:55.191 "zcopy": false, 00:14:55.191 "get_zone_info": false, 00:14:55.191 "zone_management": false, 00:14:55.191 "zone_append": false, 00:14:55.191 "compare": false, 00:14:55.191 "compare_and_write": false, 00:14:55.191 "abort": false, 00:14:55.191 "seek_hole": false, 00:14:55.191 "seek_data": false, 00:14:55.191 "copy": false, 00:14:55.191 "nvme_iov_md": false 00:14:55.191 }, 00:14:55.191 "driver_specific": { 00:14:55.191 "raid": { 00:14:55.191 "uuid": "d7073afc-9d76-4702-a72e-3a57959cdd6b", 00:14:55.191 "strip_size_kb": 64, 00:14:55.191 "state": "online", 00:14:55.191 "raid_level": "raid5f", 00:14:55.191 "superblock": true, 00:14:55.191 "num_base_bdevs": 4, 00:14:55.191 "num_base_bdevs_discovered": 4, 00:14:55.191 "num_base_bdevs_operational": 4, 00:14:55.191 "base_bdevs_list": [ 00:14:55.191 { 00:14:55.191 "name": "pt1", 00:14:55.192 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:55.192 "is_configured": true, 00:14:55.192 "data_offset": 2048, 00:14:55.192 "data_size": 63488 00:14:55.192 }, 00:14:55.192 { 00:14:55.192 "name": "pt2", 00:14:55.192 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:55.192 "is_configured": true, 00:14:55.192 "data_offset": 2048, 00:14:55.192 "data_size": 63488 00:14:55.192 }, 00:14:55.192 { 00:14:55.192 "name": "pt3", 00:14:55.192 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:55.192 "is_configured": true, 00:14:55.192 "data_offset": 2048, 00:14:55.192 "data_size": 63488 00:14:55.192 }, 00:14:55.192 { 00:14:55.192 "name": "pt4", 00:14:55.192 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:55.192 "is_configured": true, 00:14:55.192 "data_offset": 2048, 00:14:55.192 "data_size": 63488 00:14:55.192 } 00:14:55.192 ] 00:14:55.192 } 00:14:55.192 } 00:14:55.192 }' 00:14:55.192 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:55.192 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:14:55.192 pt2 00:14:55.192 pt3 00:14:55.192 pt4' 00:14:55.192 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:55.192 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:55.192 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:55.192 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:14:55.192 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:55.192 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:55.192 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:55.192 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:55.192 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:55.192 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:55.192 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:55.192 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:55.192 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:14:55.192 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:55.192 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:55.192 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:55.192 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:55.192 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:55.192 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:55.192 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:14:55.192 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:55.192 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:55.192 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:55.192 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:55.452 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:55.452 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:55.452 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:55.452 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:14:55.452 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:55.452 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:55.452 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:55.452 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:55.452 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:55.452 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:55.452 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:14:55.452 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:55.452 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:55.452 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:55.452 [2024-11-27 15:19:23.380639] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:55.452 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:55.453 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' d7073afc-9d76-4702-a72e-3a57959cdd6b '!=' d7073afc-9d76-4702-a72e-3a57959cdd6b ']' 00:14:55.453 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid5f 00:14:55.453 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:14:55.453 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:14:55.453 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:14:55.453 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:55.453 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:55.453 [2024-11-27 15:19:23.408418] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:14:55.453 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:55.453 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:55.453 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:55.453 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:55.453 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:55.453 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:55.453 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:55.453 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:55.453 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:55.453 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:55.453 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:55.453 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:55.453 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:55.453 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:55.453 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:55.453 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:55.453 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:55.453 "name": "raid_bdev1", 00:14:55.453 "uuid": "d7073afc-9d76-4702-a72e-3a57959cdd6b", 00:14:55.453 "strip_size_kb": 64, 00:14:55.453 "state": "online", 00:14:55.453 "raid_level": "raid5f", 00:14:55.453 "superblock": true, 00:14:55.453 "num_base_bdevs": 4, 00:14:55.453 "num_base_bdevs_discovered": 3, 00:14:55.453 "num_base_bdevs_operational": 3, 00:14:55.453 "base_bdevs_list": [ 00:14:55.453 { 00:14:55.453 "name": null, 00:14:55.453 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:55.453 "is_configured": false, 00:14:55.453 "data_offset": 0, 00:14:55.453 "data_size": 63488 00:14:55.453 }, 00:14:55.453 { 00:14:55.453 "name": "pt2", 00:14:55.453 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:55.453 "is_configured": true, 00:14:55.453 "data_offset": 2048, 00:14:55.453 "data_size": 63488 00:14:55.453 }, 00:14:55.453 { 00:14:55.453 "name": "pt3", 00:14:55.453 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:55.453 "is_configured": true, 00:14:55.453 "data_offset": 2048, 00:14:55.453 "data_size": 63488 00:14:55.453 }, 00:14:55.453 { 00:14:55.453 "name": "pt4", 00:14:55.453 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:55.453 "is_configured": true, 00:14:55.453 "data_offset": 2048, 00:14:55.453 "data_size": 63488 00:14:55.453 } 00:14:55.453 ] 00:14:55.453 }' 00:14:55.453 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:55.453 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.022 [2024-11-27 15:19:23.863772] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:56.022 [2024-11-27 15:19:23.863855] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:56.022 [2024-11-27 15:19:23.863966] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:56.022 [2024-11-27 15:19:23.864059] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:56.022 [2024-11-27 15:19:23.864105] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt4 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.022 [2024-11-27 15:19:23.963591] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:56.022 [2024-11-27 15:19:23.963705] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:56.022 [2024-11-27 15:19:23.963744] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:14:56.022 [2024-11-27 15:19:23.963780] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:56.022 [2024-11-27 15:19:23.966381] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:56.022 [2024-11-27 15:19:23.966468] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:56.022 [2024-11-27 15:19:23.966571] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:14:56.022 [2024-11-27 15:19:23.966630] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:56.022 pt2 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.022 15:19:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.022 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:56.022 "name": "raid_bdev1", 00:14:56.022 "uuid": "d7073afc-9d76-4702-a72e-3a57959cdd6b", 00:14:56.022 "strip_size_kb": 64, 00:14:56.022 "state": "configuring", 00:14:56.022 "raid_level": "raid5f", 00:14:56.022 "superblock": true, 00:14:56.022 "num_base_bdevs": 4, 00:14:56.022 "num_base_bdevs_discovered": 1, 00:14:56.022 "num_base_bdevs_operational": 3, 00:14:56.022 "base_bdevs_list": [ 00:14:56.022 { 00:14:56.022 "name": null, 00:14:56.022 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:56.022 "is_configured": false, 00:14:56.022 "data_offset": 2048, 00:14:56.022 "data_size": 63488 00:14:56.022 }, 00:14:56.022 { 00:14:56.022 "name": "pt2", 00:14:56.022 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:56.022 "is_configured": true, 00:14:56.022 "data_offset": 2048, 00:14:56.022 "data_size": 63488 00:14:56.022 }, 00:14:56.022 { 00:14:56.022 "name": null, 00:14:56.022 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:56.022 "is_configured": false, 00:14:56.022 "data_offset": 2048, 00:14:56.023 "data_size": 63488 00:14:56.023 }, 00:14:56.023 { 00:14:56.023 "name": null, 00:14:56.023 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:56.023 "is_configured": false, 00:14:56.023 "data_offset": 2048, 00:14:56.023 "data_size": 63488 00:14:56.023 } 00:14:56.023 ] 00:14:56.023 }' 00:14:56.023 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:56.023 15:19:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.592 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:14:56.592 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:14:56.592 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:14:56.592 15:19:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.592 15:19:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.592 [2024-11-27 15:19:24.439019] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:14:56.592 [2024-11-27 15:19:24.439168] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:56.592 [2024-11-27 15:19:24.439213] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:14:56.592 [2024-11-27 15:19:24.439257] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:56.592 [2024-11-27 15:19:24.439715] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:56.592 [2024-11-27 15:19:24.439791] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:14:56.592 [2024-11-27 15:19:24.439919] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:14:56.592 [2024-11-27 15:19:24.439986] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:56.592 pt3 00:14:56.592 15:19:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.592 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:14:56.592 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:56.592 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:56.592 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:56.592 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:56.592 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:56.592 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:56.592 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:56.592 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:56.592 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:56.592 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:56.592 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:56.592 15:19:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.592 15:19:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.592 15:19:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.592 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:56.592 "name": "raid_bdev1", 00:14:56.592 "uuid": "d7073afc-9d76-4702-a72e-3a57959cdd6b", 00:14:56.592 "strip_size_kb": 64, 00:14:56.592 "state": "configuring", 00:14:56.592 "raid_level": "raid5f", 00:14:56.592 "superblock": true, 00:14:56.592 "num_base_bdevs": 4, 00:14:56.592 "num_base_bdevs_discovered": 2, 00:14:56.592 "num_base_bdevs_operational": 3, 00:14:56.592 "base_bdevs_list": [ 00:14:56.592 { 00:14:56.592 "name": null, 00:14:56.592 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:56.592 "is_configured": false, 00:14:56.592 "data_offset": 2048, 00:14:56.592 "data_size": 63488 00:14:56.592 }, 00:14:56.592 { 00:14:56.592 "name": "pt2", 00:14:56.592 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:56.592 "is_configured": true, 00:14:56.592 "data_offset": 2048, 00:14:56.592 "data_size": 63488 00:14:56.592 }, 00:14:56.592 { 00:14:56.592 "name": "pt3", 00:14:56.592 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:56.592 "is_configured": true, 00:14:56.592 "data_offset": 2048, 00:14:56.592 "data_size": 63488 00:14:56.592 }, 00:14:56.592 { 00:14:56.592 "name": null, 00:14:56.592 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:56.592 "is_configured": false, 00:14:56.592 "data_offset": 2048, 00:14:56.592 "data_size": 63488 00:14:56.592 } 00:14:56.592 ] 00:14:56.592 }' 00:14:56.592 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:56.592 15:19:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.853 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:14:56.853 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:14:56.853 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@519 -- # i=3 00:14:56.853 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:14:56.853 15:19:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.853 15:19:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.853 [2024-11-27 15:19:24.854295] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:14:56.853 [2024-11-27 15:19:24.854410] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:56.853 [2024-11-27 15:19:24.854447] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:14:56.853 [2024-11-27 15:19:24.854482] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:56.853 [2024-11-27 15:19:24.854917] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:56.853 [2024-11-27 15:19:24.854986] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:14:56.853 [2024-11-27 15:19:24.855088] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:14:56.853 [2024-11-27 15:19:24.855143] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:14:56.853 [2024-11-27 15:19:24.855268] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:14:56.853 [2024-11-27 15:19:24.855313] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:56.853 [2024-11-27 15:19:24.855598] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:14:56.853 [2024-11-27 15:19:24.856222] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:14:56.853 [2024-11-27 15:19:24.856277] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006d00 00:14:56.853 [2024-11-27 15:19:24.856570] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:56.853 pt4 00:14:56.853 15:19:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.853 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:56.854 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:56.854 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:56.854 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:56.854 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:56.854 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:56.854 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:56.854 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:56.854 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:56.854 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:56.854 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:56.854 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:56.854 15:19:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.854 15:19:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.854 15:19:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.854 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:56.854 "name": "raid_bdev1", 00:14:56.854 "uuid": "d7073afc-9d76-4702-a72e-3a57959cdd6b", 00:14:56.854 "strip_size_kb": 64, 00:14:56.854 "state": "online", 00:14:56.854 "raid_level": "raid5f", 00:14:56.854 "superblock": true, 00:14:56.854 "num_base_bdevs": 4, 00:14:56.854 "num_base_bdevs_discovered": 3, 00:14:56.854 "num_base_bdevs_operational": 3, 00:14:56.854 "base_bdevs_list": [ 00:14:56.854 { 00:14:56.854 "name": null, 00:14:56.854 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:56.854 "is_configured": false, 00:14:56.854 "data_offset": 2048, 00:14:56.854 "data_size": 63488 00:14:56.854 }, 00:14:56.854 { 00:14:56.854 "name": "pt2", 00:14:56.854 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:56.854 "is_configured": true, 00:14:56.854 "data_offset": 2048, 00:14:56.854 "data_size": 63488 00:14:56.854 }, 00:14:56.854 { 00:14:56.854 "name": "pt3", 00:14:56.854 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:56.854 "is_configured": true, 00:14:56.854 "data_offset": 2048, 00:14:56.854 "data_size": 63488 00:14:56.854 }, 00:14:56.854 { 00:14:56.854 "name": "pt4", 00:14:56.854 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:56.854 "is_configured": true, 00:14:56.854 "data_offset": 2048, 00:14:56.854 "data_size": 63488 00:14:56.854 } 00:14:56.854 ] 00:14:56.854 }' 00:14:56.854 15:19:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:56.854 15:19:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:57.425 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:57.425 15:19:25 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:57.425 15:19:25 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:57.425 [2024-11-27 15:19:25.362468] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:57.425 [2024-11-27 15:19:25.362565] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:57.425 [2024-11-27 15:19:25.362662] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:57.425 [2024-11-27 15:19:25.362775] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:57.425 [2024-11-27 15:19:25.362826] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name raid_bdev1, state offline 00:14:57.425 15:19:25 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:57.425 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:57.425 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:14:57.425 15:19:25 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:57.425 15:19:25 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:57.425 15:19:25 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:57.425 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:14:57.425 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:14:57.425 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 4 -gt 2 ']' 00:14:57.425 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@534 -- # i=3 00:14:57.425 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt4 00:14:57.425 15:19:25 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:57.425 15:19:25 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:57.425 15:19:25 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:57.425 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:14:57.425 15:19:25 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:57.425 15:19:25 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:57.425 [2024-11-27 15:19:25.438318] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:14:57.425 [2024-11-27 15:19:25.438427] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:57.425 [2024-11-27 15:19:25.438470] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c080 00:14:57.425 [2024-11-27 15:19:25.438503] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:57.425 [2024-11-27 15:19:25.441048] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:57.425 [2024-11-27 15:19:25.441131] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:14:57.425 [2024-11-27 15:19:25.441237] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:14:57.425 [2024-11-27 15:19:25.441311] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:14:57.425 [2024-11-27 15:19:25.441497] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:14:57.426 [2024-11-27 15:19:25.441562] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:57.426 [2024-11-27 15:19:25.441611] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007080 name raid_bdev1, state configuring 00:14:57.426 [2024-11-27 15:19:25.441690] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:57.426 [2024-11-27 15:19:25.441829] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:57.426 pt1 00:14:57.426 15:19:25 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:57.426 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 4 -gt 2 ']' 00:14:57.426 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:14:57.426 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:57.426 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:57.426 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:57.426 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:57.426 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:57.426 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:57.426 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:57.426 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:57.426 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:57.426 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:57.426 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:57.426 15:19:25 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:57.426 15:19:25 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:57.426 15:19:25 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:57.426 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:57.426 "name": "raid_bdev1", 00:14:57.426 "uuid": "d7073afc-9d76-4702-a72e-3a57959cdd6b", 00:14:57.426 "strip_size_kb": 64, 00:14:57.426 "state": "configuring", 00:14:57.426 "raid_level": "raid5f", 00:14:57.426 "superblock": true, 00:14:57.426 "num_base_bdevs": 4, 00:14:57.426 "num_base_bdevs_discovered": 2, 00:14:57.426 "num_base_bdevs_operational": 3, 00:14:57.426 "base_bdevs_list": [ 00:14:57.426 { 00:14:57.426 "name": null, 00:14:57.426 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:57.426 "is_configured": false, 00:14:57.426 "data_offset": 2048, 00:14:57.426 "data_size": 63488 00:14:57.426 }, 00:14:57.426 { 00:14:57.426 "name": "pt2", 00:14:57.426 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:57.426 "is_configured": true, 00:14:57.426 "data_offset": 2048, 00:14:57.426 "data_size": 63488 00:14:57.426 }, 00:14:57.426 { 00:14:57.426 "name": "pt3", 00:14:57.426 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:57.426 "is_configured": true, 00:14:57.426 "data_offset": 2048, 00:14:57.426 "data_size": 63488 00:14:57.426 }, 00:14:57.426 { 00:14:57.426 "name": null, 00:14:57.426 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:57.426 "is_configured": false, 00:14:57.426 "data_offset": 2048, 00:14:57.426 "data_size": 63488 00:14:57.426 } 00:14:57.426 ] 00:14:57.426 }' 00:14:57.426 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:57.426 15:19:25 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:57.997 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:14:57.997 15:19:25 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:57.997 15:19:25 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:57.997 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:14:57.997 15:19:25 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:57.997 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:14:57.997 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:14:57.997 15:19:25 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:57.997 15:19:25 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:57.997 [2024-11-27 15:19:25.921517] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:14:57.997 [2024-11-27 15:19:25.921638] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:57.997 [2024-11-27 15:19:25.921683] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c680 00:14:57.997 [2024-11-27 15:19:25.921718] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:57.997 [2024-11-27 15:19:25.922152] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:57.997 [2024-11-27 15:19:25.922224] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:14:57.997 [2024-11-27 15:19:25.922328] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:14:57.997 [2024-11-27 15:19:25.922385] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:14:57.997 [2024-11-27 15:19:25.922510] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007400 00:14:57.997 [2024-11-27 15:19:25.922558] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:57.997 [2024-11-27 15:19:25.922839] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:14:57.997 [2024-11-27 15:19:25.923560] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007400 00:14:57.997 [2024-11-27 15:19:25.923625] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007400 00:14:57.997 [2024-11-27 15:19:25.923886] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:57.997 pt4 00:14:57.997 15:19:25 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:57.997 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:57.997 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:57.997 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:57.997 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:57.997 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:57.997 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:57.997 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:57.997 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:57.997 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:57.997 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:57.997 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:57.997 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:57.997 15:19:25 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:57.997 15:19:25 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:57.997 15:19:25 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:57.997 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:57.997 "name": "raid_bdev1", 00:14:57.997 "uuid": "d7073afc-9d76-4702-a72e-3a57959cdd6b", 00:14:57.997 "strip_size_kb": 64, 00:14:57.997 "state": "online", 00:14:57.997 "raid_level": "raid5f", 00:14:57.997 "superblock": true, 00:14:57.997 "num_base_bdevs": 4, 00:14:57.997 "num_base_bdevs_discovered": 3, 00:14:57.997 "num_base_bdevs_operational": 3, 00:14:57.997 "base_bdevs_list": [ 00:14:57.997 { 00:14:57.997 "name": null, 00:14:57.997 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:57.997 "is_configured": false, 00:14:57.997 "data_offset": 2048, 00:14:57.997 "data_size": 63488 00:14:57.997 }, 00:14:57.997 { 00:14:57.997 "name": "pt2", 00:14:57.997 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:57.997 "is_configured": true, 00:14:57.997 "data_offset": 2048, 00:14:57.997 "data_size": 63488 00:14:57.997 }, 00:14:57.997 { 00:14:57.997 "name": "pt3", 00:14:57.997 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:57.997 "is_configured": true, 00:14:57.997 "data_offset": 2048, 00:14:57.997 "data_size": 63488 00:14:57.997 }, 00:14:57.997 { 00:14:57.997 "name": "pt4", 00:14:57.997 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:57.997 "is_configured": true, 00:14:57.997 "data_offset": 2048, 00:14:57.997 "data_size": 63488 00:14:57.997 } 00:14:57.997 ] 00:14:57.997 }' 00:14:57.997 15:19:25 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:57.997 15:19:25 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:58.569 15:19:26 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:14:58.569 15:19:26 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:14:58.569 15:19:26 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:58.569 15:19:26 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:58.569 15:19:26 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:58.569 15:19:26 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:14:58.569 15:19:26 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:58.569 15:19:26 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:14:58.569 15:19:26 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:58.569 15:19:26 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:58.569 [2024-11-27 15:19:26.449883] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:58.569 15:19:26 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:58.569 15:19:26 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' d7073afc-9d76-4702-a72e-3a57959cdd6b '!=' d7073afc-9d76-4702-a72e-3a57959cdd6b ']' 00:14:58.569 15:19:26 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 94736 00:14:58.569 15:19:26 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 94736 ']' 00:14:58.569 15:19:26 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@958 -- # kill -0 94736 00:14:58.569 15:19:26 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@959 -- # uname 00:14:58.569 15:19:26 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:58.569 15:19:26 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 94736 00:14:58.569 killing process with pid 94736 00:14:58.569 15:19:26 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:14:58.569 15:19:26 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:14:58.569 15:19:26 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 94736' 00:14:58.569 15:19:26 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@973 -- # kill 94736 00:14:58.569 [2024-11-27 15:19:26.527374] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:58.569 [2024-11-27 15:19:26.527485] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:58.569 15:19:26 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@978 -- # wait 94736 00:14:58.569 [2024-11-27 15:19:26.527577] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:58.569 [2024-11-27 15:19:26.527590] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007400 name raid_bdev1, state offline 00:14:58.569 [2024-11-27 15:19:26.612037] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:59.139 15:19:26 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:14:59.139 00:14:59.139 real 0m7.405s 00:14:59.139 user 0m12.256s 00:14:59.139 sys 0m1.654s 00:14:59.139 15:19:26 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:59.139 ************************************ 00:14:59.139 END TEST raid5f_superblock_test 00:14:59.139 ************************************ 00:14:59.139 15:19:26 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:59.139 15:19:27 bdev_raid -- bdev/bdev_raid.sh@989 -- # '[' true = true ']' 00:14:59.139 15:19:27 bdev_raid -- bdev/bdev_raid.sh@990 -- # run_test raid5f_rebuild_test raid_rebuild_test raid5f 4 false false true 00:14:59.139 15:19:27 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:14:59.139 15:19:27 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:59.139 15:19:27 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:59.139 ************************************ 00:14:59.139 START TEST raid5f_rebuild_test 00:14:59.139 ************************************ 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid5f 4 false false true 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:14:59.139 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=95215 00:14:59.140 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:14:59.140 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 95215 00:14:59.140 15:19:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@835 -- # '[' -z 95215 ']' 00:14:59.140 15:19:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:59.140 15:19:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:59.140 15:19:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:59.140 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:59.140 15:19:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:59.140 15:19:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:59.140 [2024-11-27 15:19:27.121162] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:14:59.140 [2024-11-27 15:19:27.121752] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid95215 ] 00:14:59.140 I/O size of 3145728 is greater than zero copy threshold (65536). 00:14:59.140 Zero copy mechanism will not be used. 00:14:59.400 [2024-11-27 15:19:27.293099] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:59.400 [2024-11-27 15:19:27.334922] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:59.400 [2024-11-27 15:19:27.413571] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:59.400 [2024-11-27 15:19:27.413736] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:59.971 15:19:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:59.971 15:19:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # return 0 00:14:59.971 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:59.971 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:14:59.971 15:19:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:59.971 15:19:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:59.971 BaseBdev1_malloc 00:14:59.971 15:19:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:59.971 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:14:59.971 15:19:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:59.971 15:19:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:59.971 [2024-11-27 15:19:27.980757] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:14:59.971 [2024-11-27 15:19:27.980951] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:59.971 [2024-11-27 15:19:27.981020] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:14:59.971 [2024-11-27 15:19:27.981079] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:59.971 [2024-11-27 15:19:27.983618] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:59.971 [2024-11-27 15:19:27.983667] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:59.971 BaseBdev1 00:14:59.971 15:19:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:59.971 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:59.971 15:19:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:14:59.971 15:19:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:59.971 15:19:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:59.971 BaseBdev2_malloc 00:14:59.971 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:59.971 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:14:59.971 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:59.971 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:59.971 [2024-11-27 15:19:28.015757] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:14:59.971 [2024-11-27 15:19:28.015906] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:59.971 [2024-11-27 15:19:28.015957] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:14:59.971 [2024-11-27 15:19:28.015996] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:59.971 [2024-11-27 15:19:28.018467] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:59.971 [2024-11-27 15:19:28.018551] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:14:59.971 BaseBdev2 00:14:59.971 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:59.971 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:59.971 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:14:59.971 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:59.971 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:59.971 BaseBdev3_malloc 00:14:59.971 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:59.971 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:14:59.971 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:59.971 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:59.971 [2024-11-27 15:19:28.050651] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:14:59.971 [2024-11-27 15:19:28.050767] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:59.971 [2024-11-27 15:19:28.050815] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:14:59.971 [2024-11-27 15:19:28.050847] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:59.971 [2024-11-27 15:19:28.053267] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:59.971 [2024-11-27 15:19:28.053348] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:14:59.971 BaseBdev3 00:14:59.971 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:59.971 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:59.971 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:14:59.971 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:59.971 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:00.231 BaseBdev4_malloc 00:15:00.231 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:00.232 [2024-11-27 15:19:28.099592] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:15:00.232 [2024-11-27 15:19:28.099713] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:00.232 [2024-11-27 15:19:28.099765] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:15:00.232 [2024-11-27 15:19:28.099802] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:00.232 [2024-11-27 15:19:28.102219] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:00.232 [2024-11-27 15:19:28.102302] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:15:00.232 BaseBdev4 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:00.232 spare_malloc 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:00.232 spare_delay 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:00.232 [2024-11-27 15:19:28.151129] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:00.232 [2024-11-27 15:19:28.151322] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:00.232 [2024-11-27 15:19:28.151433] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:15:00.232 [2024-11-27 15:19:28.151538] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:00.232 [2024-11-27 15:19:28.156054] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:00.232 [2024-11-27 15:19:28.156104] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:00.232 spare 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:00.232 [2024-11-27 15:19:28.164390] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:00.232 [2024-11-27 15:19:28.166769] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:00.232 [2024-11-27 15:19:28.166902] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:00.232 [2024-11-27 15:19:28.166989] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:15:00.232 [2024-11-27 15:19:28.167128] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:15:00.232 [2024-11-27 15:19:28.167177] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:15:00.232 [2024-11-27 15:19:28.167523] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:15:00.232 [2024-11-27 15:19:28.168084] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:15:00.232 [2024-11-27 15:19:28.168149] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:15:00.232 [2024-11-27 15:19:28.168365] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:00.232 "name": "raid_bdev1", 00:15:00.232 "uuid": "e6698886-caef-4a57-856b-e2d5de09c9e8", 00:15:00.232 "strip_size_kb": 64, 00:15:00.232 "state": "online", 00:15:00.232 "raid_level": "raid5f", 00:15:00.232 "superblock": false, 00:15:00.232 "num_base_bdevs": 4, 00:15:00.232 "num_base_bdevs_discovered": 4, 00:15:00.232 "num_base_bdevs_operational": 4, 00:15:00.232 "base_bdevs_list": [ 00:15:00.232 { 00:15:00.232 "name": "BaseBdev1", 00:15:00.232 "uuid": "8c6e5f19-ab08-505f-b1f1-0d12a50d26aa", 00:15:00.232 "is_configured": true, 00:15:00.232 "data_offset": 0, 00:15:00.232 "data_size": 65536 00:15:00.232 }, 00:15:00.232 { 00:15:00.232 "name": "BaseBdev2", 00:15:00.232 "uuid": "c36c431c-0e7c-5cbd-9f03-e0f32bbe904a", 00:15:00.232 "is_configured": true, 00:15:00.232 "data_offset": 0, 00:15:00.232 "data_size": 65536 00:15:00.232 }, 00:15:00.232 { 00:15:00.232 "name": "BaseBdev3", 00:15:00.232 "uuid": "225b8d44-00e5-5bae-85a2-d1ad6dbe25b7", 00:15:00.232 "is_configured": true, 00:15:00.232 "data_offset": 0, 00:15:00.232 "data_size": 65536 00:15:00.232 }, 00:15:00.232 { 00:15:00.232 "name": "BaseBdev4", 00:15:00.232 "uuid": "751af665-2036-5cc4-ad37-f2ee4a1c3424", 00:15:00.232 "is_configured": true, 00:15:00.232 "data_offset": 0, 00:15:00.232 "data_size": 65536 00:15:00.232 } 00:15:00.232 ] 00:15:00.232 }' 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:00.232 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:00.492 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:00.492 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:15:00.492 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.492 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:00.753 [2024-11-27 15:19:28.599775] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:00.753 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.753 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=196608 00:15:00.753 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:00.753 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.753 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:00.753 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:15:00.753 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.753 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:15:00.753 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:15:00.753 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:15:00.753 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:15:00.753 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:15:00.753 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:00.753 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:15:00.753 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:00.753 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:15:00.753 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:00.753 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:15:00.753 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:00.753 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:00.753 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:15:01.013 [2024-11-27 15:19:28.875269] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:15:01.013 /dev/nbd0 00:15:01.013 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:01.013 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:01.013 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:15:01.013 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:15:01.013 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:01.013 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:01.013 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:15:01.013 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:15:01.013 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:01.013 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:01.013 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:01.013 1+0 records in 00:15:01.013 1+0 records out 00:15:01.013 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000398869 s, 10.3 MB/s 00:15:01.013 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:01.013 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:15:01.013 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:01.013 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:01.013 15:19:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:15:01.013 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:01.013 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:01.013 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:15:01.013 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@630 -- # write_unit_size=384 00:15:01.013 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@631 -- # echo 192 00:15:01.013 15:19:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=196608 count=512 oflag=direct 00:15:01.272 512+0 records in 00:15:01.272 512+0 records out 00:15:01.272 100663296 bytes (101 MB, 96 MiB) copied, 0.403054 s, 250 MB/s 00:15:01.272 15:19:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:15:01.272 15:19:29 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:01.272 15:19:29 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:15:01.272 15:19:29 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:01.272 15:19:29 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:15:01.272 15:19:29 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:01.272 15:19:29 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:01.532 15:19:29 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:01.532 [2024-11-27 15:19:29.555888] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:01.532 15:19:29 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:01.532 15:19:29 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:01.532 15:19:29 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:01.532 15:19:29 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:01.532 15:19:29 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:01.532 15:19:29 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:15:01.532 15:19:29 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:15:01.532 15:19:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:15:01.532 15:19:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:01.532 15:19:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:01.532 [2024-11-27 15:19:29.574854] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:01.532 15:19:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:01.532 15:19:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:01.532 15:19:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:01.532 15:19:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:01.532 15:19:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:01.532 15:19:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:01.532 15:19:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:01.532 15:19:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:01.532 15:19:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:01.532 15:19:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:01.532 15:19:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:01.532 15:19:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:01.532 15:19:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:01.532 15:19:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:01.532 15:19:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:01.532 15:19:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:01.532 15:19:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:01.532 "name": "raid_bdev1", 00:15:01.532 "uuid": "e6698886-caef-4a57-856b-e2d5de09c9e8", 00:15:01.532 "strip_size_kb": 64, 00:15:01.532 "state": "online", 00:15:01.532 "raid_level": "raid5f", 00:15:01.532 "superblock": false, 00:15:01.532 "num_base_bdevs": 4, 00:15:01.532 "num_base_bdevs_discovered": 3, 00:15:01.532 "num_base_bdevs_operational": 3, 00:15:01.532 "base_bdevs_list": [ 00:15:01.532 { 00:15:01.532 "name": null, 00:15:01.533 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:01.533 "is_configured": false, 00:15:01.533 "data_offset": 0, 00:15:01.533 "data_size": 65536 00:15:01.533 }, 00:15:01.533 { 00:15:01.533 "name": "BaseBdev2", 00:15:01.533 "uuid": "c36c431c-0e7c-5cbd-9f03-e0f32bbe904a", 00:15:01.533 "is_configured": true, 00:15:01.533 "data_offset": 0, 00:15:01.533 "data_size": 65536 00:15:01.533 }, 00:15:01.533 { 00:15:01.533 "name": "BaseBdev3", 00:15:01.533 "uuid": "225b8d44-00e5-5bae-85a2-d1ad6dbe25b7", 00:15:01.533 "is_configured": true, 00:15:01.533 "data_offset": 0, 00:15:01.533 "data_size": 65536 00:15:01.533 }, 00:15:01.533 { 00:15:01.533 "name": "BaseBdev4", 00:15:01.533 "uuid": "751af665-2036-5cc4-ad37-f2ee4a1c3424", 00:15:01.533 "is_configured": true, 00:15:01.533 "data_offset": 0, 00:15:01.533 "data_size": 65536 00:15:01.533 } 00:15:01.533 ] 00:15:01.533 }' 00:15:01.533 15:19:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:01.533 15:19:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.102 15:19:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:02.102 15:19:30 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.102 15:19:30 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.102 [2024-11-27 15:19:30.030116] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:02.102 [2024-11-27 15:19:30.034473] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b5b0 00:15:02.102 15:19:30 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.102 15:19:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:15:02.102 [2024-11-27 15:19:30.036740] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:03.043 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:03.043 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:03.043 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:03.043 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:03.043 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:03.043 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:03.043 15:19:31 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.043 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:03.043 15:19:31 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:03.043 15:19:31 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.043 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:03.043 "name": "raid_bdev1", 00:15:03.043 "uuid": "e6698886-caef-4a57-856b-e2d5de09c9e8", 00:15:03.043 "strip_size_kb": 64, 00:15:03.043 "state": "online", 00:15:03.043 "raid_level": "raid5f", 00:15:03.043 "superblock": false, 00:15:03.043 "num_base_bdevs": 4, 00:15:03.043 "num_base_bdevs_discovered": 4, 00:15:03.043 "num_base_bdevs_operational": 4, 00:15:03.044 "process": { 00:15:03.044 "type": "rebuild", 00:15:03.044 "target": "spare", 00:15:03.044 "progress": { 00:15:03.044 "blocks": 19200, 00:15:03.044 "percent": 9 00:15:03.044 } 00:15:03.044 }, 00:15:03.044 "base_bdevs_list": [ 00:15:03.044 { 00:15:03.044 "name": "spare", 00:15:03.044 "uuid": "69051050-c569-5a3d-abfe-792083eea04f", 00:15:03.044 "is_configured": true, 00:15:03.044 "data_offset": 0, 00:15:03.044 "data_size": 65536 00:15:03.044 }, 00:15:03.044 { 00:15:03.044 "name": "BaseBdev2", 00:15:03.044 "uuid": "c36c431c-0e7c-5cbd-9f03-e0f32bbe904a", 00:15:03.044 "is_configured": true, 00:15:03.044 "data_offset": 0, 00:15:03.044 "data_size": 65536 00:15:03.044 }, 00:15:03.044 { 00:15:03.044 "name": "BaseBdev3", 00:15:03.044 "uuid": "225b8d44-00e5-5bae-85a2-d1ad6dbe25b7", 00:15:03.044 "is_configured": true, 00:15:03.044 "data_offset": 0, 00:15:03.044 "data_size": 65536 00:15:03.044 }, 00:15:03.044 { 00:15:03.044 "name": "BaseBdev4", 00:15:03.044 "uuid": "751af665-2036-5cc4-ad37-f2ee4a1c3424", 00:15:03.044 "is_configured": true, 00:15:03.044 "data_offset": 0, 00:15:03.044 "data_size": 65536 00:15:03.044 } 00:15:03.044 ] 00:15:03.044 }' 00:15:03.044 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:03.044 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:03.044 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:03.304 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:03.304 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:03.304 15:19:31 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.304 15:19:31 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:03.304 [2024-11-27 15:19:31.193436] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:03.304 [2024-11-27 15:19:31.242677] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:03.304 [2024-11-27 15:19:31.242805] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:03.304 [2024-11-27 15:19:31.242849] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:03.304 [2024-11-27 15:19:31.242896] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:03.304 15:19:31 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.304 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:03.304 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:03.304 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:03.304 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:03.304 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:03.304 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:03.304 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:03.304 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:03.304 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:03.304 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:03.304 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:03.304 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:03.304 15:19:31 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.304 15:19:31 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:03.305 15:19:31 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.305 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:03.305 "name": "raid_bdev1", 00:15:03.305 "uuid": "e6698886-caef-4a57-856b-e2d5de09c9e8", 00:15:03.305 "strip_size_kb": 64, 00:15:03.305 "state": "online", 00:15:03.305 "raid_level": "raid5f", 00:15:03.305 "superblock": false, 00:15:03.305 "num_base_bdevs": 4, 00:15:03.305 "num_base_bdevs_discovered": 3, 00:15:03.305 "num_base_bdevs_operational": 3, 00:15:03.305 "base_bdevs_list": [ 00:15:03.305 { 00:15:03.305 "name": null, 00:15:03.305 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:03.305 "is_configured": false, 00:15:03.305 "data_offset": 0, 00:15:03.305 "data_size": 65536 00:15:03.305 }, 00:15:03.305 { 00:15:03.305 "name": "BaseBdev2", 00:15:03.305 "uuid": "c36c431c-0e7c-5cbd-9f03-e0f32bbe904a", 00:15:03.305 "is_configured": true, 00:15:03.305 "data_offset": 0, 00:15:03.305 "data_size": 65536 00:15:03.305 }, 00:15:03.305 { 00:15:03.305 "name": "BaseBdev3", 00:15:03.305 "uuid": "225b8d44-00e5-5bae-85a2-d1ad6dbe25b7", 00:15:03.305 "is_configured": true, 00:15:03.305 "data_offset": 0, 00:15:03.305 "data_size": 65536 00:15:03.305 }, 00:15:03.305 { 00:15:03.305 "name": "BaseBdev4", 00:15:03.305 "uuid": "751af665-2036-5cc4-ad37-f2ee4a1c3424", 00:15:03.305 "is_configured": true, 00:15:03.305 "data_offset": 0, 00:15:03.305 "data_size": 65536 00:15:03.305 } 00:15:03.305 ] 00:15:03.305 }' 00:15:03.305 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:03.305 15:19:31 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:03.875 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:03.875 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:03.875 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:03.875 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:03.875 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:03.875 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:03.875 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:03.875 15:19:31 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.875 15:19:31 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:03.875 15:19:31 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.875 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:03.875 "name": "raid_bdev1", 00:15:03.875 "uuid": "e6698886-caef-4a57-856b-e2d5de09c9e8", 00:15:03.875 "strip_size_kb": 64, 00:15:03.875 "state": "online", 00:15:03.875 "raid_level": "raid5f", 00:15:03.875 "superblock": false, 00:15:03.875 "num_base_bdevs": 4, 00:15:03.875 "num_base_bdevs_discovered": 3, 00:15:03.875 "num_base_bdevs_operational": 3, 00:15:03.875 "base_bdevs_list": [ 00:15:03.875 { 00:15:03.875 "name": null, 00:15:03.875 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:03.875 "is_configured": false, 00:15:03.875 "data_offset": 0, 00:15:03.875 "data_size": 65536 00:15:03.875 }, 00:15:03.875 { 00:15:03.875 "name": "BaseBdev2", 00:15:03.875 "uuid": "c36c431c-0e7c-5cbd-9f03-e0f32bbe904a", 00:15:03.875 "is_configured": true, 00:15:03.875 "data_offset": 0, 00:15:03.875 "data_size": 65536 00:15:03.875 }, 00:15:03.875 { 00:15:03.875 "name": "BaseBdev3", 00:15:03.875 "uuid": "225b8d44-00e5-5bae-85a2-d1ad6dbe25b7", 00:15:03.875 "is_configured": true, 00:15:03.875 "data_offset": 0, 00:15:03.875 "data_size": 65536 00:15:03.875 }, 00:15:03.875 { 00:15:03.875 "name": "BaseBdev4", 00:15:03.875 "uuid": "751af665-2036-5cc4-ad37-f2ee4a1c3424", 00:15:03.875 "is_configured": true, 00:15:03.875 "data_offset": 0, 00:15:03.875 "data_size": 65536 00:15:03.875 } 00:15:03.875 ] 00:15:03.875 }' 00:15:03.876 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:03.876 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:03.876 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:03.876 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:03.876 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:03.876 15:19:31 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.876 15:19:31 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:03.876 [2024-11-27 15:19:31.831841] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:03.876 [2024-11-27 15:19:31.836048] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b680 00:15:03.876 15:19:31 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.876 15:19:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:15:03.876 [2024-11-27 15:19:31.838278] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:04.816 15:19:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:04.816 15:19:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:04.816 15:19:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:04.816 15:19:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:04.816 15:19:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:04.816 15:19:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:04.816 15:19:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:04.816 15:19:32 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.816 15:19:32 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:04.816 15:19:32 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.816 15:19:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:04.816 "name": "raid_bdev1", 00:15:04.816 "uuid": "e6698886-caef-4a57-856b-e2d5de09c9e8", 00:15:04.816 "strip_size_kb": 64, 00:15:04.816 "state": "online", 00:15:04.816 "raid_level": "raid5f", 00:15:04.816 "superblock": false, 00:15:04.816 "num_base_bdevs": 4, 00:15:04.816 "num_base_bdevs_discovered": 4, 00:15:04.816 "num_base_bdevs_operational": 4, 00:15:04.816 "process": { 00:15:04.816 "type": "rebuild", 00:15:04.816 "target": "spare", 00:15:04.816 "progress": { 00:15:04.816 "blocks": 19200, 00:15:04.816 "percent": 9 00:15:04.816 } 00:15:04.816 }, 00:15:04.816 "base_bdevs_list": [ 00:15:04.816 { 00:15:04.816 "name": "spare", 00:15:04.816 "uuid": "69051050-c569-5a3d-abfe-792083eea04f", 00:15:04.816 "is_configured": true, 00:15:04.816 "data_offset": 0, 00:15:04.816 "data_size": 65536 00:15:04.816 }, 00:15:04.816 { 00:15:04.816 "name": "BaseBdev2", 00:15:04.816 "uuid": "c36c431c-0e7c-5cbd-9f03-e0f32bbe904a", 00:15:04.816 "is_configured": true, 00:15:04.816 "data_offset": 0, 00:15:04.816 "data_size": 65536 00:15:04.816 }, 00:15:04.816 { 00:15:04.816 "name": "BaseBdev3", 00:15:04.816 "uuid": "225b8d44-00e5-5bae-85a2-d1ad6dbe25b7", 00:15:04.816 "is_configured": true, 00:15:04.816 "data_offset": 0, 00:15:04.816 "data_size": 65536 00:15:04.816 }, 00:15:04.816 { 00:15:04.816 "name": "BaseBdev4", 00:15:04.816 "uuid": "751af665-2036-5cc4-ad37-f2ee4a1c3424", 00:15:04.817 "is_configured": true, 00:15:04.817 "data_offset": 0, 00:15:04.817 "data_size": 65536 00:15:04.817 } 00:15:04.817 ] 00:15:04.817 }' 00:15:04.817 15:19:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:05.077 15:19:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:05.077 15:19:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:05.077 15:19:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:05.077 15:19:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:15:05.077 15:19:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:15:05.077 15:19:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:15:05.077 15:19:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=513 00:15:05.077 15:19:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:05.077 15:19:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:05.077 15:19:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:05.077 15:19:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:05.077 15:19:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:05.077 15:19:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:05.077 15:19:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:05.078 15:19:32 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:05.078 15:19:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:05.078 15:19:32 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:05.078 15:19:33 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:05.078 15:19:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:05.078 "name": "raid_bdev1", 00:15:05.078 "uuid": "e6698886-caef-4a57-856b-e2d5de09c9e8", 00:15:05.078 "strip_size_kb": 64, 00:15:05.078 "state": "online", 00:15:05.078 "raid_level": "raid5f", 00:15:05.078 "superblock": false, 00:15:05.078 "num_base_bdevs": 4, 00:15:05.078 "num_base_bdevs_discovered": 4, 00:15:05.078 "num_base_bdevs_operational": 4, 00:15:05.078 "process": { 00:15:05.078 "type": "rebuild", 00:15:05.078 "target": "spare", 00:15:05.078 "progress": { 00:15:05.078 "blocks": 21120, 00:15:05.078 "percent": 10 00:15:05.078 } 00:15:05.078 }, 00:15:05.078 "base_bdevs_list": [ 00:15:05.078 { 00:15:05.078 "name": "spare", 00:15:05.078 "uuid": "69051050-c569-5a3d-abfe-792083eea04f", 00:15:05.078 "is_configured": true, 00:15:05.078 "data_offset": 0, 00:15:05.078 "data_size": 65536 00:15:05.078 }, 00:15:05.078 { 00:15:05.078 "name": "BaseBdev2", 00:15:05.078 "uuid": "c36c431c-0e7c-5cbd-9f03-e0f32bbe904a", 00:15:05.078 "is_configured": true, 00:15:05.078 "data_offset": 0, 00:15:05.078 "data_size": 65536 00:15:05.078 }, 00:15:05.078 { 00:15:05.078 "name": "BaseBdev3", 00:15:05.078 "uuid": "225b8d44-00e5-5bae-85a2-d1ad6dbe25b7", 00:15:05.078 "is_configured": true, 00:15:05.078 "data_offset": 0, 00:15:05.078 "data_size": 65536 00:15:05.078 }, 00:15:05.078 { 00:15:05.078 "name": "BaseBdev4", 00:15:05.078 "uuid": "751af665-2036-5cc4-ad37-f2ee4a1c3424", 00:15:05.078 "is_configured": true, 00:15:05.078 "data_offset": 0, 00:15:05.078 "data_size": 65536 00:15:05.078 } 00:15:05.078 ] 00:15:05.078 }' 00:15:05.078 15:19:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:05.078 15:19:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:05.078 15:19:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:05.078 15:19:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:05.078 15:19:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:06.020 15:19:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:06.020 15:19:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:06.020 15:19:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:06.020 15:19:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:06.020 15:19:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:06.020 15:19:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:06.020 15:19:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:06.020 15:19:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:06.020 15:19:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:06.020 15:19:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:06.280 15:19:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:06.280 15:19:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:06.280 "name": "raid_bdev1", 00:15:06.280 "uuid": "e6698886-caef-4a57-856b-e2d5de09c9e8", 00:15:06.280 "strip_size_kb": 64, 00:15:06.280 "state": "online", 00:15:06.280 "raid_level": "raid5f", 00:15:06.280 "superblock": false, 00:15:06.280 "num_base_bdevs": 4, 00:15:06.280 "num_base_bdevs_discovered": 4, 00:15:06.280 "num_base_bdevs_operational": 4, 00:15:06.280 "process": { 00:15:06.280 "type": "rebuild", 00:15:06.280 "target": "spare", 00:15:06.280 "progress": { 00:15:06.280 "blocks": 42240, 00:15:06.280 "percent": 21 00:15:06.280 } 00:15:06.280 }, 00:15:06.280 "base_bdevs_list": [ 00:15:06.280 { 00:15:06.280 "name": "spare", 00:15:06.280 "uuid": "69051050-c569-5a3d-abfe-792083eea04f", 00:15:06.280 "is_configured": true, 00:15:06.280 "data_offset": 0, 00:15:06.280 "data_size": 65536 00:15:06.280 }, 00:15:06.280 { 00:15:06.280 "name": "BaseBdev2", 00:15:06.280 "uuid": "c36c431c-0e7c-5cbd-9f03-e0f32bbe904a", 00:15:06.280 "is_configured": true, 00:15:06.280 "data_offset": 0, 00:15:06.280 "data_size": 65536 00:15:06.280 }, 00:15:06.280 { 00:15:06.280 "name": "BaseBdev3", 00:15:06.280 "uuid": "225b8d44-00e5-5bae-85a2-d1ad6dbe25b7", 00:15:06.280 "is_configured": true, 00:15:06.280 "data_offset": 0, 00:15:06.280 "data_size": 65536 00:15:06.280 }, 00:15:06.280 { 00:15:06.280 "name": "BaseBdev4", 00:15:06.280 "uuid": "751af665-2036-5cc4-ad37-f2ee4a1c3424", 00:15:06.280 "is_configured": true, 00:15:06.280 "data_offset": 0, 00:15:06.280 "data_size": 65536 00:15:06.280 } 00:15:06.280 ] 00:15:06.280 }' 00:15:06.280 15:19:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:06.281 15:19:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:06.281 15:19:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:06.281 15:19:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:06.281 15:19:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:07.221 15:19:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:07.221 15:19:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:07.221 15:19:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:07.221 15:19:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:07.221 15:19:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:07.221 15:19:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:07.221 15:19:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:07.221 15:19:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:07.221 15:19:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:07.221 15:19:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:07.221 15:19:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:07.221 15:19:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:07.221 "name": "raid_bdev1", 00:15:07.221 "uuid": "e6698886-caef-4a57-856b-e2d5de09c9e8", 00:15:07.221 "strip_size_kb": 64, 00:15:07.221 "state": "online", 00:15:07.221 "raid_level": "raid5f", 00:15:07.221 "superblock": false, 00:15:07.221 "num_base_bdevs": 4, 00:15:07.221 "num_base_bdevs_discovered": 4, 00:15:07.221 "num_base_bdevs_operational": 4, 00:15:07.221 "process": { 00:15:07.221 "type": "rebuild", 00:15:07.221 "target": "spare", 00:15:07.221 "progress": { 00:15:07.221 "blocks": 65280, 00:15:07.221 "percent": 33 00:15:07.222 } 00:15:07.222 }, 00:15:07.222 "base_bdevs_list": [ 00:15:07.222 { 00:15:07.222 "name": "spare", 00:15:07.222 "uuid": "69051050-c569-5a3d-abfe-792083eea04f", 00:15:07.222 "is_configured": true, 00:15:07.222 "data_offset": 0, 00:15:07.222 "data_size": 65536 00:15:07.222 }, 00:15:07.222 { 00:15:07.222 "name": "BaseBdev2", 00:15:07.222 "uuid": "c36c431c-0e7c-5cbd-9f03-e0f32bbe904a", 00:15:07.222 "is_configured": true, 00:15:07.222 "data_offset": 0, 00:15:07.222 "data_size": 65536 00:15:07.222 }, 00:15:07.222 { 00:15:07.222 "name": "BaseBdev3", 00:15:07.222 "uuid": "225b8d44-00e5-5bae-85a2-d1ad6dbe25b7", 00:15:07.222 "is_configured": true, 00:15:07.222 "data_offset": 0, 00:15:07.222 "data_size": 65536 00:15:07.222 }, 00:15:07.222 { 00:15:07.222 "name": "BaseBdev4", 00:15:07.222 "uuid": "751af665-2036-5cc4-ad37-f2ee4a1c3424", 00:15:07.222 "is_configured": true, 00:15:07.222 "data_offset": 0, 00:15:07.222 "data_size": 65536 00:15:07.222 } 00:15:07.222 ] 00:15:07.222 }' 00:15:07.222 15:19:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:07.482 15:19:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:07.482 15:19:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:07.482 15:19:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:07.482 15:19:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:08.423 15:19:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:08.423 15:19:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:08.423 15:19:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:08.423 15:19:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:08.423 15:19:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:08.423 15:19:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:08.423 15:19:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:08.423 15:19:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:08.423 15:19:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:08.423 15:19:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:08.423 15:19:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:08.423 15:19:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:08.423 "name": "raid_bdev1", 00:15:08.423 "uuid": "e6698886-caef-4a57-856b-e2d5de09c9e8", 00:15:08.423 "strip_size_kb": 64, 00:15:08.423 "state": "online", 00:15:08.423 "raid_level": "raid5f", 00:15:08.423 "superblock": false, 00:15:08.423 "num_base_bdevs": 4, 00:15:08.423 "num_base_bdevs_discovered": 4, 00:15:08.423 "num_base_bdevs_operational": 4, 00:15:08.423 "process": { 00:15:08.423 "type": "rebuild", 00:15:08.423 "target": "spare", 00:15:08.423 "progress": { 00:15:08.423 "blocks": 86400, 00:15:08.423 "percent": 43 00:15:08.423 } 00:15:08.423 }, 00:15:08.423 "base_bdevs_list": [ 00:15:08.423 { 00:15:08.423 "name": "spare", 00:15:08.423 "uuid": "69051050-c569-5a3d-abfe-792083eea04f", 00:15:08.423 "is_configured": true, 00:15:08.423 "data_offset": 0, 00:15:08.423 "data_size": 65536 00:15:08.423 }, 00:15:08.423 { 00:15:08.423 "name": "BaseBdev2", 00:15:08.423 "uuid": "c36c431c-0e7c-5cbd-9f03-e0f32bbe904a", 00:15:08.423 "is_configured": true, 00:15:08.423 "data_offset": 0, 00:15:08.423 "data_size": 65536 00:15:08.423 }, 00:15:08.423 { 00:15:08.423 "name": "BaseBdev3", 00:15:08.423 "uuid": "225b8d44-00e5-5bae-85a2-d1ad6dbe25b7", 00:15:08.423 "is_configured": true, 00:15:08.423 "data_offset": 0, 00:15:08.423 "data_size": 65536 00:15:08.423 }, 00:15:08.423 { 00:15:08.423 "name": "BaseBdev4", 00:15:08.423 "uuid": "751af665-2036-5cc4-ad37-f2ee4a1c3424", 00:15:08.423 "is_configured": true, 00:15:08.423 "data_offset": 0, 00:15:08.423 "data_size": 65536 00:15:08.423 } 00:15:08.423 ] 00:15:08.423 }' 00:15:08.423 15:19:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:08.423 15:19:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:08.423 15:19:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:08.683 15:19:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:08.683 15:19:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:09.622 15:19:37 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:09.622 15:19:37 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:09.622 15:19:37 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:09.622 15:19:37 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:09.622 15:19:37 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:09.622 15:19:37 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:09.622 15:19:37 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:09.622 15:19:37 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.622 15:19:37 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:09.622 15:19:37 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:09.622 15:19:37 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.622 15:19:37 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:09.622 "name": "raid_bdev1", 00:15:09.622 "uuid": "e6698886-caef-4a57-856b-e2d5de09c9e8", 00:15:09.622 "strip_size_kb": 64, 00:15:09.622 "state": "online", 00:15:09.622 "raid_level": "raid5f", 00:15:09.622 "superblock": false, 00:15:09.622 "num_base_bdevs": 4, 00:15:09.622 "num_base_bdevs_discovered": 4, 00:15:09.622 "num_base_bdevs_operational": 4, 00:15:09.622 "process": { 00:15:09.622 "type": "rebuild", 00:15:09.622 "target": "spare", 00:15:09.622 "progress": { 00:15:09.622 "blocks": 107520, 00:15:09.622 "percent": 54 00:15:09.622 } 00:15:09.622 }, 00:15:09.622 "base_bdevs_list": [ 00:15:09.622 { 00:15:09.622 "name": "spare", 00:15:09.622 "uuid": "69051050-c569-5a3d-abfe-792083eea04f", 00:15:09.622 "is_configured": true, 00:15:09.622 "data_offset": 0, 00:15:09.622 "data_size": 65536 00:15:09.622 }, 00:15:09.622 { 00:15:09.622 "name": "BaseBdev2", 00:15:09.622 "uuid": "c36c431c-0e7c-5cbd-9f03-e0f32bbe904a", 00:15:09.622 "is_configured": true, 00:15:09.622 "data_offset": 0, 00:15:09.622 "data_size": 65536 00:15:09.622 }, 00:15:09.622 { 00:15:09.622 "name": "BaseBdev3", 00:15:09.622 "uuid": "225b8d44-00e5-5bae-85a2-d1ad6dbe25b7", 00:15:09.622 "is_configured": true, 00:15:09.622 "data_offset": 0, 00:15:09.622 "data_size": 65536 00:15:09.622 }, 00:15:09.622 { 00:15:09.622 "name": "BaseBdev4", 00:15:09.622 "uuid": "751af665-2036-5cc4-ad37-f2ee4a1c3424", 00:15:09.622 "is_configured": true, 00:15:09.622 "data_offset": 0, 00:15:09.622 "data_size": 65536 00:15:09.622 } 00:15:09.622 ] 00:15:09.622 }' 00:15:09.622 15:19:37 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:09.622 15:19:37 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:09.622 15:19:37 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:09.622 15:19:37 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:09.622 15:19:37 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:11.004 15:19:38 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:11.004 15:19:38 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:11.004 15:19:38 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:11.004 15:19:38 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:11.004 15:19:38 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:11.004 15:19:38 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:11.004 15:19:38 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:11.004 15:19:38 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:11.004 15:19:38 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:11.004 15:19:38 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:11.004 15:19:38 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:11.004 15:19:38 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:11.004 "name": "raid_bdev1", 00:15:11.004 "uuid": "e6698886-caef-4a57-856b-e2d5de09c9e8", 00:15:11.004 "strip_size_kb": 64, 00:15:11.004 "state": "online", 00:15:11.004 "raid_level": "raid5f", 00:15:11.004 "superblock": false, 00:15:11.004 "num_base_bdevs": 4, 00:15:11.004 "num_base_bdevs_discovered": 4, 00:15:11.004 "num_base_bdevs_operational": 4, 00:15:11.004 "process": { 00:15:11.004 "type": "rebuild", 00:15:11.004 "target": "spare", 00:15:11.004 "progress": { 00:15:11.004 "blocks": 130560, 00:15:11.004 "percent": 66 00:15:11.004 } 00:15:11.004 }, 00:15:11.004 "base_bdevs_list": [ 00:15:11.004 { 00:15:11.004 "name": "spare", 00:15:11.004 "uuid": "69051050-c569-5a3d-abfe-792083eea04f", 00:15:11.004 "is_configured": true, 00:15:11.004 "data_offset": 0, 00:15:11.004 "data_size": 65536 00:15:11.004 }, 00:15:11.004 { 00:15:11.004 "name": "BaseBdev2", 00:15:11.004 "uuid": "c36c431c-0e7c-5cbd-9f03-e0f32bbe904a", 00:15:11.004 "is_configured": true, 00:15:11.004 "data_offset": 0, 00:15:11.004 "data_size": 65536 00:15:11.004 }, 00:15:11.004 { 00:15:11.004 "name": "BaseBdev3", 00:15:11.004 "uuid": "225b8d44-00e5-5bae-85a2-d1ad6dbe25b7", 00:15:11.004 "is_configured": true, 00:15:11.004 "data_offset": 0, 00:15:11.004 "data_size": 65536 00:15:11.004 }, 00:15:11.004 { 00:15:11.004 "name": "BaseBdev4", 00:15:11.004 "uuid": "751af665-2036-5cc4-ad37-f2ee4a1c3424", 00:15:11.004 "is_configured": true, 00:15:11.004 "data_offset": 0, 00:15:11.004 "data_size": 65536 00:15:11.004 } 00:15:11.004 ] 00:15:11.004 }' 00:15:11.004 15:19:38 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:11.004 15:19:38 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:11.004 15:19:38 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:11.004 15:19:38 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:11.004 15:19:38 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:11.950 15:19:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:11.950 15:19:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:11.950 15:19:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:11.950 15:19:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:11.950 15:19:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:11.950 15:19:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:11.950 15:19:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:11.950 15:19:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:11.950 15:19:39 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:11.950 15:19:39 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:11.950 15:19:39 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:11.950 15:19:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:11.950 "name": "raid_bdev1", 00:15:11.950 "uuid": "e6698886-caef-4a57-856b-e2d5de09c9e8", 00:15:11.950 "strip_size_kb": 64, 00:15:11.950 "state": "online", 00:15:11.950 "raid_level": "raid5f", 00:15:11.950 "superblock": false, 00:15:11.950 "num_base_bdevs": 4, 00:15:11.950 "num_base_bdevs_discovered": 4, 00:15:11.950 "num_base_bdevs_operational": 4, 00:15:11.950 "process": { 00:15:11.950 "type": "rebuild", 00:15:11.950 "target": "spare", 00:15:11.950 "progress": { 00:15:11.950 "blocks": 151680, 00:15:11.950 "percent": 77 00:15:11.950 } 00:15:11.950 }, 00:15:11.950 "base_bdevs_list": [ 00:15:11.950 { 00:15:11.950 "name": "spare", 00:15:11.950 "uuid": "69051050-c569-5a3d-abfe-792083eea04f", 00:15:11.950 "is_configured": true, 00:15:11.950 "data_offset": 0, 00:15:11.950 "data_size": 65536 00:15:11.950 }, 00:15:11.950 { 00:15:11.950 "name": "BaseBdev2", 00:15:11.950 "uuid": "c36c431c-0e7c-5cbd-9f03-e0f32bbe904a", 00:15:11.950 "is_configured": true, 00:15:11.950 "data_offset": 0, 00:15:11.950 "data_size": 65536 00:15:11.950 }, 00:15:11.950 { 00:15:11.950 "name": "BaseBdev3", 00:15:11.950 "uuid": "225b8d44-00e5-5bae-85a2-d1ad6dbe25b7", 00:15:11.950 "is_configured": true, 00:15:11.950 "data_offset": 0, 00:15:11.950 "data_size": 65536 00:15:11.950 }, 00:15:11.950 { 00:15:11.950 "name": "BaseBdev4", 00:15:11.950 "uuid": "751af665-2036-5cc4-ad37-f2ee4a1c3424", 00:15:11.950 "is_configured": true, 00:15:11.950 "data_offset": 0, 00:15:11.950 "data_size": 65536 00:15:11.950 } 00:15:11.950 ] 00:15:11.950 }' 00:15:11.950 15:19:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:11.951 15:19:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:11.951 15:19:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:11.951 15:19:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:11.951 15:19:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:12.960 15:19:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:12.960 15:19:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:12.960 15:19:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:12.960 15:19:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:12.960 15:19:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:12.960 15:19:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:12.960 15:19:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:12.960 15:19:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:12.960 15:19:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:12.960 15:19:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:12.960 15:19:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:12.960 15:19:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:12.960 "name": "raid_bdev1", 00:15:12.960 "uuid": "e6698886-caef-4a57-856b-e2d5de09c9e8", 00:15:12.960 "strip_size_kb": 64, 00:15:12.960 "state": "online", 00:15:12.960 "raid_level": "raid5f", 00:15:12.960 "superblock": false, 00:15:12.960 "num_base_bdevs": 4, 00:15:12.960 "num_base_bdevs_discovered": 4, 00:15:12.960 "num_base_bdevs_operational": 4, 00:15:12.960 "process": { 00:15:12.960 "type": "rebuild", 00:15:12.960 "target": "spare", 00:15:12.960 "progress": { 00:15:12.960 "blocks": 172800, 00:15:12.960 "percent": 87 00:15:12.960 } 00:15:12.960 }, 00:15:12.960 "base_bdevs_list": [ 00:15:12.960 { 00:15:12.960 "name": "spare", 00:15:12.960 "uuid": "69051050-c569-5a3d-abfe-792083eea04f", 00:15:12.960 "is_configured": true, 00:15:12.960 "data_offset": 0, 00:15:12.960 "data_size": 65536 00:15:12.960 }, 00:15:12.960 { 00:15:12.960 "name": "BaseBdev2", 00:15:12.960 "uuid": "c36c431c-0e7c-5cbd-9f03-e0f32bbe904a", 00:15:12.960 "is_configured": true, 00:15:12.960 "data_offset": 0, 00:15:12.960 "data_size": 65536 00:15:12.960 }, 00:15:12.960 { 00:15:12.960 "name": "BaseBdev3", 00:15:12.960 "uuid": "225b8d44-00e5-5bae-85a2-d1ad6dbe25b7", 00:15:12.960 "is_configured": true, 00:15:12.960 "data_offset": 0, 00:15:12.960 "data_size": 65536 00:15:12.960 }, 00:15:12.960 { 00:15:12.960 "name": "BaseBdev4", 00:15:12.960 "uuid": "751af665-2036-5cc4-ad37-f2ee4a1c3424", 00:15:12.960 "is_configured": true, 00:15:12.960 "data_offset": 0, 00:15:12.960 "data_size": 65536 00:15:12.960 } 00:15:12.960 ] 00:15:12.961 }' 00:15:12.961 15:19:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:12.961 15:19:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:12.961 15:19:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:13.221 15:19:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:13.221 15:19:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:14.162 15:19:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:14.162 15:19:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:14.162 15:19:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:14.162 15:19:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:14.162 15:19:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:14.162 15:19:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:14.162 15:19:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:14.162 15:19:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:14.162 15:19:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:14.162 15:19:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:14.162 15:19:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:14.162 15:19:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:14.162 "name": "raid_bdev1", 00:15:14.162 "uuid": "e6698886-caef-4a57-856b-e2d5de09c9e8", 00:15:14.162 "strip_size_kb": 64, 00:15:14.162 "state": "online", 00:15:14.162 "raid_level": "raid5f", 00:15:14.162 "superblock": false, 00:15:14.162 "num_base_bdevs": 4, 00:15:14.162 "num_base_bdevs_discovered": 4, 00:15:14.162 "num_base_bdevs_operational": 4, 00:15:14.162 "process": { 00:15:14.162 "type": "rebuild", 00:15:14.162 "target": "spare", 00:15:14.162 "progress": { 00:15:14.162 "blocks": 193920, 00:15:14.162 "percent": 98 00:15:14.162 } 00:15:14.162 }, 00:15:14.162 "base_bdevs_list": [ 00:15:14.162 { 00:15:14.162 "name": "spare", 00:15:14.162 "uuid": "69051050-c569-5a3d-abfe-792083eea04f", 00:15:14.162 "is_configured": true, 00:15:14.162 "data_offset": 0, 00:15:14.162 "data_size": 65536 00:15:14.162 }, 00:15:14.162 { 00:15:14.162 "name": "BaseBdev2", 00:15:14.162 "uuid": "c36c431c-0e7c-5cbd-9f03-e0f32bbe904a", 00:15:14.162 "is_configured": true, 00:15:14.162 "data_offset": 0, 00:15:14.162 "data_size": 65536 00:15:14.162 }, 00:15:14.162 { 00:15:14.162 "name": "BaseBdev3", 00:15:14.162 "uuid": "225b8d44-00e5-5bae-85a2-d1ad6dbe25b7", 00:15:14.162 "is_configured": true, 00:15:14.162 "data_offset": 0, 00:15:14.162 "data_size": 65536 00:15:14.162 }, 00:15:14.162 { 00:15:14.162 "name": "BaseBdev4", 00:15:14.162 "uuid": "751af665-2036-5cc4-ad37-f2ee4a1c3424", 00:15:14.162 "is_configured": true, 00:15:14.162 "data_offset": 0, 00:15:14.162 "data_size": 65536 00:15:14.162 } 00:15:14.162 ] 00:15:14.162 }' 00:15:14.162 15:19:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:14.162 15:19:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:14.162 15:19:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:14.162 [2024-11-27 15:19:42.193306] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:15:14.162 [2024-11-27 15:19:42.193486] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:15:14.162 [2024-11-27 15:19:42.193572] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:14.162 15:19:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:14.162 15:19:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:15.102 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:15.102 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:15.102 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:15.102 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:15.102 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:15.102 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:15.363 "name": "raid_bdev1", 00:15:15.363 "uuid": "e6698886-caef-4a57-856b-e2d5de09c9e8", 00:15:15.363 "strip_size_kb": 64, 00:15:15.363 "state": "online", 00:15:15.363 "raid_level": "raid5f", 00:15:15.363 "superblock": false, 00:15:15.363 "num_base_bdevs": 4, 00:15:15.363 "num_base_bdevs_discovered": 4, 00:15:15.363 "num_base_bdevs_operational": 4, 00:15:15.363 "base_bdevs_list": [ 00:15:15.363 { 00:15:15.363 "name": "spare", 00:15:15.363 "uuid": "69051050-c569-5a3d-abfe-792083eea04f", 00:15:15.363 "is_configured": true, 00:15:15.363 "data_offset": 0, 00:15:15.363 "data_size": 65536 00:15:15.363 }, 00:15:15.363 { 00:15:15.363 "name": "BaseBdev2", 00:15:15.363 "uuid": "c36c431c-0e7c-5cbd-9f03-e0f32bbe904a", 00:15:15.363 "is_configured": true, 00:15:15.363 "data_offset": 0, 00:15:15.363 "data_size": 65536 00:15:15.363 }, 00:15:15.363 { 00:15:15.363 "name": "BaseBdev3", 00:15:15.363 "uuid": "225b8d44-00e5-5bae-85a2-d1ad6dbe25b7", 00:15:15.363 "is_configured": true, 00:15:15.363 "data_offset": 0, 00:15:15.363 "data_size": 65536 00:15:15.363 }, 00:15:15.363 { 00:15:15.363 "name": "BaseBdev4", 00:15:15.363 "uuid": "751af665-2036-5cc4-ad37-f2ee4a1c3424", 00:15:15.363 "is_configured": true, 00:15:15.363 "data_offset": 0, 00:15:15.363 "data_size": 65536 00:15:15.363 } 00:15:15.363 ] 00:15:15.363 }' 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:15.363 "name": "raid_bdev1", 00:15:15.363 "uuid": "e6698886-caef-4a57-856b-e2d5de09c9e8", 00:15:15.363 "strip_size_kb": 64, 00:15:15.363 "state": "online", 00:15:15.363 "raid_level": "raid5f", 00:15:15.363 "superblock": false, 00:15:15.363 "num_base_bdevs": 4, 00:15:15.363 "num_base_bdevs_discovered": 4, 00:15:15.363 "num_base_bdevs_operational": 4, 00:15:15.363 "base_bdevs_list": [ 00:15:15.363 { 00:15:15.363 "name": "spare", 00:15:15.363 "uuid": "69051050-c569-5a3d-abfe-792083eea04f", 00:15:15.363 "is_configured": true, 00:15:15.363 "data_offset": 0, 00:15:15.363 "data_size": 65536 00:15:15.363 }, 00:15:15.363 { 00:15:15.363 "name": "BaseBdev2", 00:15:15.363 "uuid": "c36c431c-0e7c-5cbd-9f03-e0f32bbe904a", 00:15:15.363 "is_configured": true, 00:15:15.363 "data_offset": 0, 00:15:15.363 "data_size": 65536 00:15:15.363 }, 00:15:15.363 { 00:15:15.363 "name": "BaseBdev3", 00:15:15.363 "uuid": "225b8d44-00e5-5bae-85a2-d1ad6dbe25b7", 00:15:15.363 "is_configured": true, 00:15:15.363 "data_offset": 0, 00:15:15.363 "data_size": 65536 00:15:15.363 }, 00:15:15.363 { 00:15:15.363 "name": "BaseBdev4", 00:15:15.363 "uuid": "751af665-2036-5cc4-ad37-f2ee4a1c3424", 00:15:15.363 "is_configured": true, 00:15:15.363 "data_offset": 0, 00:15:15.363 "data_size": 65536 00:15:15.363 } 00:15:15.363 ] 00:15:15.363 }' 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:15.363 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:15.623 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:15.623 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:15.623 15:19:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:15.623 15:19:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:15.623 15:19:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:15.623 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:15.623 "name": "raid_bdev1", 00:15:15.623 "uuid": "e6698886-caef-4a57-856b-e2d5de09c9e8", 00:15:15.623 "strip_size_kb": 64, 00:15:15.623 "state": "online", 00:15:15.623 "raid_level": "raid5f", 00:15:15.623 "superblock": false, 00:15:15.623 "num_base_bdevs": 4, 00:15:15.623 "num_base_bdevs_discovered": 4, 00:15:15.623 "num_base_bdevs_operational": 4, 00:15:15.623 "base_bdevs_list": [ 00:15:15.623 { 00:15:15.623 "name": "spare", 00:15:15.623 "uuid": "69051050-c569-5a3d-abfe-792083eea04f", 00:15:15.623 "is_configured": true, 00:15:15.623 "data_offset": 0, 00:15:15.623 "data_size": 65536 00:15:15.623 }, 00:15:15.623 { 00:15:15.623 "name": "BaseBdev2", 00:15:15.623 "uuid": "c36c431c-0e7c-5cbd-9f03-e0f32bbe904a", 00:15:15.623 "is_configured": true, 00:15:15.623 "data_offset": 0, 00:15:15.623 "data_size": 65536 00:15:15.623 }, 00:15:15.623 { 00:15:15.623 "name": "BaseBdev3", 00:15:15.623 "uuid": "225b8d44-00e5-5bae-85a2-d1ad6dbe25b7", 00:15:15.623 "is_configured": true, 00:15:15.623 "data_offset": 0, 00:15:15.623 "data_size": 65536 00:15:15.624 }, 00:15:15.624 { 00:15:15.624 "name": "BaseBdev4", 00:15:15.624 "uuid": "751af665-2036-5cc4-ad37-f2ee4a1c3424", 00:15:15.624 "is_configured": true, 00:15:15.624 "data_offset": 0, 00:15:15.624 "data_size": 65536 00:15:15.624 } 00:15:15.624 ] 00:15:15.624 }' 00:15:15.624 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:15.624 15:19:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:15.884 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:15.884 15:19:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:15.884 15:19:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:15.884 [2024-11-27 15:19:43.936440] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:15.884 [2024-11-27 15:19:43.936582] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:15.884 [2024-11-27 15:19:43.936724] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:15.884 [2024-11-27 15:19:43.936849] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:15.884 [2024-11-27 15:19:43.936933] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:15:15.884 15:19:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:15.884 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:15.884 15:19:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:15.884 15:19:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:15.884 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:15:15.884 15:19:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:15.884 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:15:15.884 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:15:15.884 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:15:15.884 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:15:15.884 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:15.884 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:15:15.884 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:15.884 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:15.884 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:15.884 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:15:15.884 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:15.884 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:15.884 15:19:43 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:15:16.144 /dev/nbd0 00:15:16.144 15:19:44 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:16.144 15:19:44 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:16.144 15:19:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:15:16.144 15:19:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:15:16.144 15:19:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:16.144 15:19:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:16.144 15:19:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:15:16.144 15:19:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:15:16.144 15:19:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:16.144 15:19:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:16.144 15:19:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:16.144 1+0 records in 00:15:16.144 1+0 records out 00:15:16.144 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000532704 s, 7.7 MB/s 00:15:16.144 15:19:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:16.144 15:19:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:15:16.144 15:19:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:16.144 15:19:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:16.144 15:19:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:15:16.144 15:19:44 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:16.144 15:19:44 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:16.144 15:19:44 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:15:16.405 /dev/nbd1 00:15:16.405 15:19:44 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:15:16.405 15:19:44 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:15:16.405 15:19:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:15:16.405 15:19:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:15:16.405 15:19:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:16.405 15:19:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:16.405 15:19:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:15:16.405 15:19:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:15:16.405 15:19:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:16.405 15:19:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:16.405 15:19:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:16.405 1+0 records in 00:15:16.405 1+0 records out 00:15:16.405 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.0002886 s, 14.2 MB/s 00:15:16.405 15:19:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:16.405 15:19:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:15:16.405 15:19:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:16.405 15:19:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:16.405 15:19:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:15:16.405 15:19:44 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:16.405 15:19:44 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:16.405 15:19:44 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:15:16.665 15:19:44 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:15:16.665 15:19:44 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:16.665 15:19:44 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:16.665 15:19:44 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:16.665 15:19:44 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:15:16.665 15:19:44 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:16.665 15:19:44 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:16.925 15:19:44 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:16.925 15:19:44 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:16.925 15:19:44 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:16.925 15:19:44 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:16.925 15:19:44 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:16.925 15:19:44 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:16.925 15:19:44 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:15:16.925 15:19:44 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:15:16.925 15:19:44 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:16.925 15:19:44 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:15:16.925 15:19:44 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:15:16.925 15:19:45 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:15:16.925 15:19:45 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:15:16.925 15:19:45 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:16.925 15:19:45 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:16.925 15:19:45 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:15:16.925 15:19:45 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:15:16.925 15:19:45 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:15:16.925 15:19:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:15:16.925 15:19:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 95215 00:15:16.925 15:19:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@954 -- # '[' -z 95215 ']' 00:15:16.925 15:19:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@958 -- # kill -0 95215 00:15:16.925 15:19:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@959 -- # uname 00:15:16.925 15:19:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:16.925 15:19:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 95215 00:15:17.186 15:19:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:17.186 15:19:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:17.186 15:19:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 95215' 00:15:17.186 killing process with pid 95215 00:15:17.186 15:19:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@973 -- # kill 95215 00:15:17.186 Received shutdown signal, test time was about 60.000000 seconds 00:15:17.186 00:15:17.186 Latency(us) 00:15:17.186 [2024-11-27T15:19:45.293Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:15:17.186 [2024-11-27T15:19:45.293Z] =================================================================================================================== 00:15:17.186 [2024-11-27T15:19:45.293Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:15:17.186 [2024-11-27 15:19:45.056344] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:17.186 15:19:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@978 -- # wait 95215 00:15:17.186 [2024-11-27 15:19:45.149398] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:15:17.447 00:15:17.447 real 0m18.438s 00:15:17.447 user 0m22.162s 00:15:17.447 sys 0m2.271s 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:17.447 ************************************ 00:15:17.447 END TEST raid5f_rebuild_test 00:15:17.447 ************************************ 00:15:17.447 15:19:45 bdev_raid -- bdev/bdev_raid.sh@991 -- # run_test raid5f_rebuild_test_sb raid_rebuild_test raid5f 4 true false true 00:15:17.447 15:19:45 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:15:17.447 15:19:45 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:17.447 15:19:45 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:17.447 ************************************ 00:15:17.447 START TEST raid5f_rebuild_test_sb 00:15:17.447 ************************************ 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid5f 4 true false true 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=95720 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 95720 00:15:17.447 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:15:17.448 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@835 -- # '[' -z 95720 ']' 00:15:17.448 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:17.448 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:17.448 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:17.448 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:17.448 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:17.708 15:19:45 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:17.708 I/O size of 3145728 is greater than zero copy threshold (65536). 00:15:17.708 Zero copy mechanism will not be used. 00:15:17.708 [2024-11-27 15:19:45.630057] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:15:17.708 [2024-11-27 15:19:45.630185] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid95720 ] 00:15:17.708 [2024-11-27 15:19:45.799146] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:17.968 [2024-11-27 15:19:45.837441] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:17.968 [2024-11-27 15:19:45.913019] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:17.968 [2024-11-27 15:19:45.913067] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # return 0 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:18.539 BaseBdev1_malloc 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:18.539 [2024-11-27 15:19:46.481541] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:18.539 [2024-11-27 15:19:46.481616] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:18.539 [2024-11-27 15:19:46.481653] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:18.539 [2024-11-27 15:19:46.481673] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:18.539 [2024-11-27 15:19:46.484043] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:18.539 [2024-11-27 15:19:46.484080] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:18.539 BaseBdev1 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:18.539 BaseBdev2_malloc 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:18.539 [2024-11-27 15:19:46.515966] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:15:18.539 [2024-11-27 15:19:46.516024] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:18.539 [2024-11-27 15:19:46.516049] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:15:18.539 [2024-11-27 15:19:46.516059] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:18.539 [2024-11-27 15:19:46.518360] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:18.539 [2024-11-27 15:19:46.518398] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:15:18.539 BaseBdev2 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:18.539 BaseBdev3_malloc 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:18.539 [2024-11-27 15:19:46.550388] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:15:18.539 [2024-11-27 15:19:46.550445] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:18.539 [2024-11-27 15:19:46.550474] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:15:18.539 [2024-11-27 15:19:46.550484] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:18.539 [2024-11-27 15:19:46.552769] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:18.539 [2024-11-27 15:19:46.552809] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:15:18.539 BaseBdev3 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:18.539 BaseBdev4_malloc 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:18.539 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:18.539 [2024-11-27 15:19:46.602093] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:15:18.539 [2024-11-27 15:19:46.602184] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:18.539 [2024-11-27 15:19:46.602232] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:15:18.540 [2024-11-27 15:19:46.602251] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:18.540 [2024-11-27 15:19:46.606009] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:18.540 [2024-11-27 15:19:46.606063] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:15:18.540 BaseBdev4 00:15:18.540 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:18.540 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:15:18.540 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:18.540 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:18.540 spare_malloc 00:15:18.540 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:18.540 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:15:18.540 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:18.540 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:18.540 spare_delay 00:15:18.540 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:18.540 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:18.540 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:18.540 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:18.801 [2024-11-27 15:19:46.648832] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:18.801 [2024-11-27 15:19:46.648883] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:18.801 [2024-11-27 15:19:46.648930] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:15:18.801 [2024-11-27 15:19:46.648942] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:18.801 [2024-11-27 15:19:46.651291] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:18.801 [2024-11-27 15:19:46.651330] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:18.801 spare 00:15:18.801 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:18.801 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:15:18.801 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:18.801 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:18.801 [2024-11-27 15:19:46.660887] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:18.801 [2024-11-27 15:19:46.662953] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:18.801 [2024-11-27 15:19:46.663031] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:18.801 [2024-11-27 15:19:46.663076] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:15:18.801 [2024-11-27 15:19:46.663266] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:15:18.801 [2024-11-27 15:19:46.663287] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:18.801 [2024-11-27 15:19:46.663576] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:15:18.801 [2024-11-27 15:19:46.664074] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:15:18.801 [2024-11-27 15:19:46.664100] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:15:18.801 [2024-11-27 15:19:46.664234] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:18.801 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:18.801 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:15:18.801 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:18.801 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:18.801 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:18.801 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:18.801 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:18.801 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:18.801 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:18.801 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:18.801 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:18.801 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:18.801 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:18.801 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:18.801 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:18.801 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:18.801 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:18.801 "name": "raid_bdev1", 00:15:18.801 "uuid": "e95708c2-d41d-45cb-9efa-c4b368f39949", 00:15:18.801 "strip_size_kb": 64, 00:15:18.801 "state": "online", 00:15:18.801 "raid_level": "raid5f", 00:15:18.801 "superblock": true, 00:15:18.801 "num_base_bdevs": 4, 00:15:18.801 "num_base_bdevs_discovered": 4, 00:15:18.801 "num_base_bdevs_operational": 4, 00:15:18.801 "base_bdevs_list": [ 00:15:18.801 { 00:15:18.801 "name": "BaseBdev1", 00:15:18.801 "uuid": "5b8f1f84-f2aa-572e-b5ef-97b5f938d5e2", 00:15:18.801 "is_configured": true, 00:15:18.801 "data_offset": 2048, 00:15:18.801 "data_size": 63488 00:15:18.801 }, 00:15:18.801 { 00:15:18.801 "name": "BaseBdev2", 00:15:18.801 "uuid": "0a632a98-aa29-5e4f-8d5c-732f5f7d1023", 00:15:18.801 "is_configured": true, 00:15:18.801 "data_offset": 2048, 00:15:18.801 "data_size": 63488 00:15:18.801 }, 00:15:18.801 { 00:15:18.801 "name": "BaseBdev3", 00:15:18.801 "uuid": "b3ec0800-b7f3-543b-baa8-c25a0e9115d1", 00:15:18.801 "is_configured": true, 00:15:18.801 "data_offset": 2048, 00:15:18.801 "data_size": 63488 00:15:18.801 }, 00:15:18.801 { 00:15:18.802 "name": "BaseBdev4", 00:15:18.802 "uuid": "315998c8-8084-57d1-bce6-0a4b3d0b594d", 00:15:18.802 "is_configured": true, 00:15:18.802 "data_offset": 2048, 00:15:18.802 "data_size": 63488 00:15:18.802 } 00:15:18.802 ] 00:15:18.802 }' 00:15:18.802 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:18.802 15:19:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:19.062 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:19.062 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:15:19.062 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:19.062 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:19.062 [2024-11-27 15:19:47.138831] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:19.062 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:19.322 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=190464 00:15:19.322 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:15:19.322 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:19.322 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:19.322 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:19.322 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:19.322 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:15:19.322 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:15:19.322 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:15:19.322 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:15:19.322 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:15:19.322 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:19.322 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:15:19.322 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:19.322 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:15:19.322 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:19.322 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:15:19.322 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:19.322 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:19.322 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:15:19.323 [2024-11-27 15:19:47.422160] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:15:19.581 /dev/nbd0 00:15:19.581 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:19.581 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:19.581 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:15:19.581 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:15:19.581 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:19.581 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:19.581 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:15:19.581 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:15:19.581 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:19.581 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:19.581 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:19.581 1+0 records in 00:15:19.581 1+0 records out 00:15:19.581 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000339263 s, 12.1 MB/s 00:15:19.581 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:19.581 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:15:19.581 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:19.581 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:19.581 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:15:19.581 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:19.581 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:19.581 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:15:19.581 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@630 -- # write_unit_size=384 00:15:19.581 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@631 -- # echo 192 00:15:19.581 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=196608 count=496 oflag=direct 00:15:19.839 496+0 records in 00:15:19.839 496+0 records out 00:15:19.840 97517568 bytes (98 MB, 93 MiB) copied, 0.39857 s, 245 MB/s 00:15:19.840 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:15:19.840 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:19.840 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:15:19.840 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:19.840 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:15:19.840 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:19.840 15:19:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:20.100 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:20.100 [2024-11-27 15:19:48.102721] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:20.100 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:20.100 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:20.100 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:20.100 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:20.100 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:20.100 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:15:20.100 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:15:20.100 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:15:20.100 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:20.100 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:20.100 [2024-11-27 15:19:48.114781] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:20.100 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:20.100 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:20.100 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:20.100 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:20.100 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:20.100 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:20.100 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:20.100 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:20.100 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:20.100 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:20.100 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:20.100 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:20.100 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:20.100 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:20.100 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:20.100 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:20.100 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:20.100 "name": "raid_bdev1", 00:15:20.100 "uuid": "e95708c2-d41d-45cb-9efa-c4b368f39949", 00:15:20.100 "strip_size_kb": 64, 00:15:20.100 "state": "online", 00:15:20.100 "raid_level": "raid5f", 00:15:20.100 "superblock": true, 00:15:20.100 "num_base_bdevs": 4, 00:15:20.100 "num_base_bdevs_discovered": 3, 00:15:20.100 "num_base_bdevs_operational": 3, 00:15:20.100 "base_bdevs_list": [ 00:15:20.100 { 00:15:20.100 "name": null, 00:15:20.100 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:20.100 "is_configured": false, 00:15:20.100 "data_offset": 0, 00:15:20.100 "data_size": 63488 00:15:20.100 }, 00:15:20.100 { 00:15:20.100 "name": "BaseBdev2", 00:15:20.100 "uuid": "0a632a98-aa29-5e4f-8d5c-732f5f7d1023", 00:15:20.100 "is_configured": true, 00:15:20.100 "data_offset": 2048, 00:15:20.100 "data_size": 63488 00:15:20.100 }, 00:15:20.100 { 00:15:20.100 "name": "BaseBdev3", 00:15:20.100 "uuid": "b3ec0800-b7f3-543b-baa8-c25a0e9115d1", 00:15:20.100 "is_configured": true, 00:15:20.100 "data_offset": 2048, 00:15:20.100 "data_size": 63488 00:15:20.100 }, 00:15:20.100 { 00:15:20.100 "name": "BaseBdev4", 00:15:20.100 "uuid": "315998c8-8084-57d1-bce6-0a4b3d0b594d", 00:15:20.100 "is_configured": true, 00:15:20.100 "data_offset": 2048, 00:15:20.100 "data_size": 63488 00:15:20.100 } 00:15:20.100 ] 00:15:20.100 }' 00:15:20.100 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:20.100 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:20.669 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:20.669 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:20.669 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:20.669 [2024-11-27 15:19:48.582026] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:20.669 [2024-11-27 15:19:48.586485] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002a8b0 00:15:20.669 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:20.669 15:19:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:15:20.669 [2024-11-27 15:19:48.588867] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:21.611 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:21.611 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:21.611 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:21.611 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:21.611 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:21.611 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:21.611 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:21.611 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:21.611 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:21.611 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:21.611 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:21.611 "name": "raid_bdev1", 00:15:21.611 "uuid": "e95708c2-d41d-45cb-9efa-c4b368f39949", 00:15:21.611 "strip_size_kb": 64, 00:15:21.611 "state": "online", 00:15:21.611 "raid_level": "raid5f", 00:15:21.611 "superblock": true, 00:15:21.611 "num_base_bdevs": 4, 00:15:21.611 "num_base_bdevs_discovered": 4, 00:15:21.611 "num_base_bdevs_operational": 4, 00:15:21.611 "process": { 00:15:21.611 "type": "rebuild", 00:15:21.611 "target": "spare", 00:15:21.611 "progress": { 00:15:21.611 "blocks": 19200, 00:15:21.611 "percent": 10 00:15:21.611 } 00:15:21.611 }, 00:15:21.611 "base_bdevs_list": [ 00:15:21.611 { 00:15:21.611 "name": "spare", 00:15:21.611 "uuid": "4f8f62cb-4a1f-59cd-a325-c384b3bd334f", 00:15:21.611 "is_configured": true, 00:15:21.611 "data_offset": 2048, 00:15:21.611 "data_size": 63488 00:15:21.611 }, 00:15:21.611 { 00:15:21.611 "name": "BaseBdev2", 00:15:21.611 "uuid": "0a632a98-aa29-5e4f-8d5c-732f5f7d1023", 00:15:21.611 "is_configured": true, 00:15:21.611 "data_offset": 2048, 00:15:21.611 "data_size": 63488 00:15:21.611 }, 00:15:21.611 { 00:15:21.611 "name": "BaseBdev3", 00:15:21.611 "uuid": "b3ec0800-b7f3-543b-baa8-c25a0e9115d1", 00:15:21.611 "is_configured": true, 00:15:21.611 "data_offset": 2048, 00:15:21.611 "data_size": 63488 00:15:21.611 }, 00:15:21.611 { 00:15:21.611 "name": "BaseBdev4", 00:15:21.611 "uuid": "315998c8-8084-57d1-bce6-0a4b3d0b594d", 00:15:21.611 "is_configured": true, 00:15:21.611 "data_offset": 2048, 00:15:21.611 "data_size": 63488 00:15:21.611 } 00:15:21.611 ] 00:15:21.611 }' 00:15:21.611 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:21.611 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:21.611 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:21.871 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:21.871 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:21.871 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:21.871 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:21.871 [2024-11-27 15:19:49.752914] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:21.871 [2024-11-27 15:19:49.795369] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:21.871 [2024-11-27 15:19:49.795446] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:21.871 [2024-11-27 15:19:49.795469] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:21.871 [2024-11-27 15:19:49.795477] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:21.871 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:21.871 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:21.871 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:21.871 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:21.871 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:21.871 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:21.871 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:21.871 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:21.871 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:21.871 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:21.871 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:21.871 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:21.871 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:21.871 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:21.871 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:21.871 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:21.871 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:21.871 "name": "raid_bdev1", 00:15:21.871 "uuid": "e95708c2-d41d-45cb-9efa-c4b368f39949", 00:15:21.871 "strip_size_kb": 64, 00:15:21.871 "state": "online", 00:15:21.871 "raid_level": "raid5f", 00:15:21.871 "superblock": true, 00:15:21.871 "num_base_bdevs": 4, 00:15:21.871 "num_base_bdevs_discovered": 3, 00:15:21.871 "num_base_bdevs_operational": 3, 00:15:21.871 "base_bdevs_list": [ 00:15:21.871 { 00:15:21.871 "name": null, 00:15:21.871 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:21.871 "is_configured": false, 00:15:21.871 "data_offset": 0, 00:15:21.871 "data_size": 63488 00:15:21.871 }, 00:15:21.871 { 00:15:21.871 "name": "BaseBdev2", 00:15:21.871 "uuid": "0a632a98-aa29-5e4f-8d5c-732f5f7d1023", 00:15:21.871 "is_configured": true, 00:15:21.871 "data_offset": 2048, 00:15:21.871 "data_size": 63488 00:15:21.871 }, 00:15:21.871 { 00:15:21.871 "name": "BaseBdev3", 00:15:21.871 "uuid": "b3ec0800-b7f3-543b-baa8-c25a0e9115d1", 00:15:21.871 "is_configured": true, 00:15:21.871 "data_offset": 2048, 00:15:21.871 "data_size": 63488 00:15:21.871 }, 00:15:21.871 { 00:15:21.871 "name": "BaseBdev4", 00:15:21.871 "uuid": "315998c8-8084-57d1-bce6-0a4b3d0b594d", 00:15:21.871 "is_configured": true, 00:15:21.871 "data_offset": 2048, 00:15:21.871 "data_size": 63488 00:15:21.871 } 00:15:21.871 ] 00:15:21.871 }' 00:15:21.871 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:21.871 15:19:49 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:22.440 15:19:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:22.440 15:19:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:22.440 15:19:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:22.440 15:19:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:22.440 15:19:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:22.440 15:19:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:22.440 15:19:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:22.440 15:19:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:22.440 15:19:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:22.440 15:19:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:22.440 15:19:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:22.440 "name": "raid_bdev1", 00:15:22.440 "uuid": "e95708c2-d41d-45cb-9efa-c4b368f39949", 00:15:22.440 "strip_size_kb": 64, 00:15:22.440 "state": "online", 00:15:22.440 "raid_level": "raid5f", 00:15:22.441 "superblock": true, 00:15:22.441 "num_base_bdevs": 4, 00:15:22.441 "num_base_bdevs_discovered": 3, 00:15:22.441 "num_base_bdevs_operational": 3, 00:15:22.441 "base_bdevs_list": [ 00:15:22.441 { 00:15:22.441 "name": null, 00:15:22.441 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:22.441 "is_configured": false, 00:15:22.441 "data_offset": 0, 00:15:22.441 "data_size": 63488 00:15:22.441 }, 00:15:22.441 { 00:15:22.441 "name": "BaseBdev2", 00:15:22.441 "uuid": "0a632a98-aa29-5e4f-8d5c-732f5f7d1023", 00:15:22.441 "is_configured": true, 00:15:22.441 "data_offset": 2048, 00:15:22.441 "data_size": 63488 00:15:22.441 }, 00:15:22.441 { 00:15:22.441 "name": "BaseBdev3", 00:15:22.441 "uuid": "b3ec0800-b7f3-543b-baa8-c25a0e9115d1", 00:15:22.441 "is_configured": true, 00:15:22.441 "data_offset": 2048, 00:15:22.441 "data_size": 63488 00:15:22.441 }, 00:15:22.441 { 00:15:22.441 "name": "BaseBdev4", 00:15:22.441 "uuid": "315998c8-8084-57d1-bce6-0a4b3d0b594d", 00:15:22.441 "is_configured": true, 00:15:22.441 "data_offset": 2048, 00:15:22.441 "data_size": 63488 00:15:22.441 } 00:15:22.441 ] 00:15:22.441 }' 00:15:22.441 15:19:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:22.441 15:19:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:22.441 15:19:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:22.441 15:19:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:22.441 15:19:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:22.441 15:19:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:22.441 15:19:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:22.441 [2024-11-27 15:19:50.376499] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:22.441 [2024-11-27 15:19:50.380873] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002a980 00:15:22.441 15:19:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:22.441 15:19:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:15:22.441 [2024-11-27 15:19:50.383098] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:23.382 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:23.383 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:23.383 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:23.383 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:23.383 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:23.383 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:23.383 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:23.383 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:23.383 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:23.383 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:23.383 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:23.383 "name": "raid_bdev1", 00:15:23.383 "uuid": "e95708c2-d41d-45cb-9efa-c4b368f39949", 00:15:23.383 "strip_size_kb": 64, 00:15:23.383 "state": "online", 00:15:23.383 "raid_level": "raid5f", 00:15:23.383 "superblock": true, 00:15:23.383 "num_base_bdevs": 4, 00:15:23.383 "num_base_bdevs_discovered": 4, 00:15:23.383 "num_base_bdevs_operational": 4, 00:15:23.383 "process": { 00:15:23.383 "type": "rebuild", 00:15:23.383 "target": "spare", 00:15:23.383 "progress": { 00:15:23.383 "blocks": 19200, 00:15:23.383 "percent": 10 00:15:23.383 } 00:15:23.383 }, 00:15:23.383 "base_bdevs_list": [ 00:15:23.383 { 00:15:23.383 "name": "spare", 00:15:23.383 "uuid": "4f8f62cb-4a1f-59cd-a325-c384b3bd334f", 00:15:23.383 "is_configured": true, 00:15:23.383 "data_offset": 2048, 00:15:23.383 "data_size": 63488 00:15:23.383 }, 00:15:23.383 { 00:15:23.383 "name": "BaseBdev2", 00:15:23.383 "uuid": "0a632a98-aa29-5e4f-8d5c-732f5f7d1023", 00:15:23.383 "is_configured": true, 00:15:23.383 "data_offset": 2048, 00:15:23.383 "data_size": 63488 00:15:23.383 }, 00:15:23.383 { 00:15:23.383 "name": "BaseBdev3", 00:15:23.383 "uuid": "b3ec0800-b7f3-543b-baa8-c25a0e9115d1", 00:15:23.383 "is_configured": true, 00:15:23.383 "data_offset": 2048, 00:15:23.383 "data_size": 63488 00:15:23.383 }, 00:15:23.383 { 00:15:23.383 "name": "BaseBdev4", 00:15:23.383 "uuid": "315998c8-8084-57d1-bce6-0a4b3d0b594d", 00:15:23.383 "is_configured": true, 00:15:23.383 "data_offset": 2048, 00:15:23.383 "data_size": 63488 00:15:23.383 } 00:15:23.383 ] 00:15:23.383 }' 00:15:23.383 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:23.644 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:23.644 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:23.644 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:23.644 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:15:23.644 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:15:23.644 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:15:23.644 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:15:23.644 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:15:23.644 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=532 00:15:23.644 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:23.644 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:23.644 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:23.644 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:23.644 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:23.644 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:23.644 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:23.644 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:23.644 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:23.644 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:23.644 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:23.644 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:23.644 "name": "raid_bdev1", 00:15:23.644 "uuid": "e95708c2-d41d-45cb-9efa-c4b368f39949", 00:15:23.644 "strip_size_kb": 64, 00:15:23.644 "state": "online", 00:15:23.644 "raid_level": "raid5f", 00:15:23.644 "superblock": true, 00:15:23.644 "num_base_bdevs": 4, 00:15:23.644 "num_base_bdevs_discovered": 4, 00:15:23.644 "num_base_bdevs_operational": 4, 00:15:23.644 "process": { 00:15:23.644 "type": "rebuild", 00:15:23.644 "target": "spare", 00:15:23.644 "progress": { 00:15:23.644 "blocks": 21120, 00:15:23.644 "percent": 11 00:15:23.644 } 00:15:23.644 }, 00:15:23.644 "base_bdevs_list": [ 00:15:23.644 { 00:15:23.644 "name": "spare", 00:15:23.644 "uuid": "4f8f62cb-4a1f-59cd-a325-c384b3bd334f", 00:15:23.644 "is_configured": true, 00:15:23.644 "data_offset": 2048, 00:15:23.644 "data_size": 63488 00:15:23.644 }, 00:15:23.644 { 00:15:23.644 "name": "BaseBdev2", 00:15:23.644 "uuid": "0a632a98-aa29-5e4f-8d5c-732f5f7d1023", 00:15:23.644 "is_configured": true, 00:15:23.644 "data_offset": 2048, 00:15:23.644 "data_size": 63488 00:15:23.644 }, 00:15:23.644 { 00:15:23.644 "name": "BaseBdev3", 00:15:23.644 "uuid": "b3ec0800-b7f3-543b-baa8-c25a0e9115d1", 00:15:23.644 "is_configured": true, 00:15:23.644 "data_offset": 2048, 00:15:23.644 "data_size": 63488 00:15:23.644 }, 00:15:23.644 { 00:15:23.644 "name": "BaseBdev4", 00:15:23.644 "uuid": "315998c8-8084-57d1-bce6-0a4b3d0b594d", 00:15:23.644 "is_configured": true, 00:15:23.644 "data_offset": 2048, 00:15:23.644 "data_size": 63488 00:15:23.644 } 00:15:23.644 ] 00:15:23.644 }' 00:15:23.644 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:23.644 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:23.644 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:23.644 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:23.644 15:19:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:25.026 15:19:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:25.026 15:19:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:25.026 15:19:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:25.026 15:19:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:25.026 15:19:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:25.026 15:19:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:25.026 15:19:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:25.026 15:19:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:25.026 15:19:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:25.026 15:19:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:25.026 15:19:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:25.026 15:19:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:25.026 "name": "raid_bdev1", 00:15:25.026 "uuid": "e95708c2-d41d-45cb-9efa-c4b368f39949", 00:15:25.026 "strip_size_kb": 64, 00:15:25.026 "state": "online", 00:15:25.026 "raid_level": "raid5f", 00:15:25.026 "superblock": true, 00:15:25.026 "num_base_bdevs": 4, 00:15:25.026 "num_base_bdevs_discovered": 4, 00:15:25.026 "num_base_bdevs_operational": 4, 00:15:25.026 "process": { 00:15:25.026 "type": "rebuild", 00:15:25.026 "target": "spare", 00:15:25.026 "progress": { 00:15:25.026 "blocks": 44160, 00:15:25.026 "percent": 23 00:15:25.026 } 00:15:25.026 }, 00:15:25.026 "base_bdevs_list": [ 00:15:25.026 { 00:15:25.026 "name": "spare", 00:15:25.026 "uuid": "4f8f62cb-4a1f-59cd-a325-c384b3bd334f", 00:15:25.026 "is_configured": true, 00:15:25.026 "data_offset": 2048, 00:15:25.026 "data_size": 63488 00:15:25.026 }, 00:15:25.026 { 00:15:25.026 "name": "BaseBdev2", 00:15:25.026 "uuid": "0a632a98-aa29-5e4f-8d5c-732f5f7d1023", 00:15:25.026 "is_configured": true, 00:15:25.026 "data_offset": 2048, 00:15:25.026 "data_size": 63488 00:15:25.026 }, 00:15:25.026 { 00:15:25.026 "name": "BaseBdev3", 00:15:25.026 "uuid": "b3ec0800-b7f3-543b-baa8-c25a0e9115d1", 00:15:25.026 "is_configured": true, 00:15:25.026 "data_offset": 2048, 00:15:25.026 "data_size": 63488 00:15:25.026 }, 00:15:25.026 { 00:15:25.026 "name": "BaseBdev4", 00:15:25.026 "uuid": "315998c8-8084-57d1-bce6-0a4b3d0b594d", 00:15:25.026 "is_configured": true, 00:15:25.026 "data_offset": 2048, 00:15:25.026 "data_size": 63488 00:15:25.026 } 00:15:25.026 ] 00:15:25.026 }' 00:15:25.026 15:19:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:25.026 15:19:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:25.026 15:19:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:25.026 15:19:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:25.026 15:19:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:25.965 15:19:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:25.965 15:19:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:25.965 15:19:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:25.965 15:19:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:25.965 15:19:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:25.965 15:19:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:25.965 15:19:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:25.965 15:19:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:25.965 15:19:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:25.965 15:19:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:25.965 15:19:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:25.965 15:19:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:25.965 "name": "raid_bdev1", 00:15:25.965 "uuid": "e95708c2-d41d-45cb-9efa-c4b368f39949", 00:15:25.965 "strip_size_kb": 64, 00:15:25.965 "state": "online", 00:15:25.966 "raid_level": "raid5f", 00:15:25.966 "superblock": true, 00:15:25.966 "num_base_bdevs": 4, 00:15:25.966 "num_base_bdevs_discovered": 4, 00:15:25.966 "num_base_bdevs_operational": 4, 00:15:25.966 "process": { 00:15:25.966 "type": "rebuild", 00:15:25.966 "target": "spare", 00:15:25.966 "progress": { 00:15:25.966 "blocks": 65280, 00:15:25.966 "percent": 34 00:15:25.966 } 00:15:25.966 }, 00:15:25.966 "base_bdevs_list": [ 00:15:25.966 { 00:15:25.966 "name": "spare", 00:15:25.966 "uuid": "4f8f62cb-4a1f-59cd-a325-c384b3bd334f", 00:15:25.966 "is_configured": true, 00:15:25.966 "data_offset": 2048, 00:15:25.966 "data_size": 63488 00:15:25.966 }, 00:15:25.966 { 00:15:25.966 "name": "BaseBdev2", 00:15:25.966 "uuid": "0a632a98-aa29-5e4f-8d5c-732f5f7d1023", 00:15:25.966 "is_configured": true, 00:15:25.966 "data_offset": 2048, 00:15:25.966 "data_size": 63488 00:15:25.966 }, 00:15:25.966 { 00:15:25.966 "name": "BaseBdev3", 00:15:25.966 "uuid": "b3ec0800-b7f3-543b-baa8-c25a0e9115d1", 00:15:25.966 "is_configured": true, 00:15:25.966 "data_offset": 2048, 00:15:25.966 "data_size": 63488 00:15:25.966 }, 00:15:25.966 { 00:15:25.966 "name": "BaseBdev4", 00:15:25.966 "uuid": "315998c8-8084-57d1-bce6-0a4b3d0b594d", 00:15:25.966 "is_configured": true, 00:15:25.966 "data_offset": 2048, 00:15:25.966 "data_size": 63488 00:15:25.966 } 00:15:25.966 ] 00:15:25.966 }' 00:15:25.966 15:19:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:25.966 15:19:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:25.966 15:19:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:25.966 15:19:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:25.966 15:19:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:26.926 15:19:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:26.926 15:19:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:26.926 15:19:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:26.926 15:19:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:26.926 15:19:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:26.926 15:19:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:26.926 15:19:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:26.926 15:19:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:26.926 15:19:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:26.926 15:19:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:27.200 15:19:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:27.200 15:19:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:27.200 "name": "raid_bdev1", 00:15:27.200 "uuid": "e95708c2-d41d-45cb-9efa-c4b368f39949", 00:15:27.200 "strip_size_kb": 64, 00:15:27.200 "state": "online", 00:15:27.200 "raid_level": "raid5f", 00:15:27.200 "superblock": true, 00:15:27.200 "num_base_bdevs": 4, 00:15:27.200 "num_base_bdevs_discovered": 4, 00:15:27.200 "num_base_bdevs_operational": 4, 00:15:27.200 "process": { 00:15:27.200 "type": "rebuild", 00:15:27.200 "target": "spare", 00:15:27.200 "progress": { 00:15:27.200 "blocks": 88320, 00:15:27.200 "percent": 46 00:15:27.200 } 00:15:27.200 }, 00:15:27.200 "base_bdevs_list": [ 00:15:27.200 { 00:15:27.200 "name": "spare", 00:15:27.200 "uuid": "4f8f62cb-4a1f-59cd-a325-c384b3bd334f", 00:15:27.200 "is_configured": true, 00:15:27.200 "data_offset": 2048, 00:15:27.200 "data_size": 63488 00:15:27.200 }, 00:15:27.200 { 00:15:27.200 "name": "BaseBdev2", 00:15:27.200 "uuid": "0a632a98-aa29-5e4f-8d5c-732f5f7d1023", 00:15:27.200 "is_configured": true, 00:15:27.200 "data_offset": 2048, 00:15:27.200 "data_size": 63488 00:15:27.200 }, 00:15:27.200 { 00:15:27.200 "name": "BaseBdev3", 00:15:27.200 "uuid": "b3ec0800-b7f3-543b-baa8-c25a0e9115d1", 00:15:27.200 "is_configured": true, 00:15:27.200 "data_offset": 2048, 00:15:27.200 "data_size": 63488 00:15:27.200 }, 00:15:27.200 { 00:15:27.200 "name": "BaseBdev4", 00:15:27.200 "uuid": "315998c8-8084-57d1-bce6-0a4b3d0b594d", 00:15:27.200 "is_configured": true, 00:15:27.200 "data_offset": 2048, 00:15:27.200 "data_size": 63488 00:15:27.200 } 00:15:27.200 ] 00:15:27.200 }' 00:15:27.200 15:19:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:27.200 15:19:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:27.200 15:19:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:27.200 15:19:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:27.200 15:19:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:28.138 15:19:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:28.138 15:19:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:28.138 15:19:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:28.138 15:19:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:28.138 15:19:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:28.138 15:19:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:28.138 15:19:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:28.138 15:19:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:28.138 15:19:56 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:28.138 15:19:56 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:28.138 15:19:56 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:28.138 15:19:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:28.138 "name": "raid_bdev1", 00:15:28.138 "uuid": "e95708c2-d41d-45cb-9efa-c4b368f39949", 00:15:28.138 "strip_size_kb": 64, 00:15:28.138 "state": "online", 00:15:28.138 "raid_level": "raid5f", 00:15:28.138 "superblock": true, 00:15:28.138 "num_base_bdevs": 4, 00:15:28.138 "num_base_bdevs_discovered": 4, 00:15:28.138 "num_base_bdevs_operational": 4, 00:15:28.138 "process": { 00:15:28.138 "type": "rebuild", 00:15:28.138 "target": "spare", 00:15:28.138 "progress": { 00:15:28.138 "blocks": 109440, 00:15:28.138 "percent": 57 00:15:28.138 } 00:15:28.138 }, 00:15:28.138 "base_bdevs_list": [ 00:15:28.138 { 00:15:28.138 "name": "spare", 00:15:28.138 "uuid": "4f8f62cb-4a1f-59cd-a325-c384b3bd334f", 00:15:28.138 "is_configured": true, 00:15:28.138 "data_offset": 2048, 00:15:28.138 "data_size": 63488 00:15:28.138 }, 00:15:28.138 { 00:15:28.138 "name": "BaseBdev2", 00:15:28.138 "uuid": "0a632a98-aa29-5e4f-8d5c-732f5f7d1023", 00:15:28.138 "is_configured": true, 00:15:28.138 "data_offset": 2048, 00:15:28.138 "data_size": 63488 00:15:28.138 }, 00:15:28.138 { 00:15:28.138 "name": "BaseBdev3", 00:15:28.138 "uuid": "b3ec0800-b7f3-543b-baa8-c25a0e9115d1", 00:15:28.138 "is_configured": true, 00:15:28.138 "data_offset": 2048, 00:15:28.138 "data_size": 63488 00:15:28.138 }, 00:15:28.138 { 00:15:28.138 "name": "BaseBdev4", 00:15:28.138 "uuid": "315998c8-8084-57d1-bce6-0a4b3d0b594d", 00:15:28.138 "is_configured": true, 00:15:28.138 "data_offset": 2048, 00:15:28.138 "data_size": 63488 00:15:28.138 } 00:15:28.138 ] 00:15:28.138 }' 00:15:28.138 15:19:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:28.397 15:19:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:28.397 15:19:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:28.397 15:19:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:28.397 15:19:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:29.335 15:19:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:29.335 15:19:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:29.335 15:19:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:29.335 15:19:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:29.335 15:19:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:29.335 15:19:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:29.335 15:19:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:29.335 15:19:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:29.335 15:19:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:29.335 15:19:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:29.335 15:19:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:29.335 15:19:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:29.335 "name": "raid_bdev1", 00:15:29.335 "uuid": "e95708c2-d41d-45cb-9efa-c4b368f39949", 00:15:29.335 "strip_size_kb": 64, 00:15:29.335 "state": "online", 00:15:29.335 "raid_level": "raid5f", 00:15:29.335 "superblock": true, 00:15:29.335 "num_base_bdevs": 4, 00:15:29.335 "num_base_bdevs_discovered": 4, 00:15:29.335 "num_base_bdevs_operational": 4, 00:15:29.335 "process": { 00:15:29.335 "type": "rebuild", 00:15:29.335 "target": "spare", 00:15:29.335 "progress": { 00:15:29.335 "blocks": 132480, 00:15:29.335 "percent": 69 00:15:29.335 } 00:15:29.335 }, 00:15:29.335 "base_bdevs_list": [ 00:15:29.335 { 00:15:29.335 "name": "spare", 00:15:29.335 "uuid": "4f8f62cb-4a1f-59cd-a325-c384b3bd334f", 00:15:29.335 "is_configured": true, 00:15:29.335 "data_offset": 2048, 00:15:29.335 "data_size": 63488 00:15:29.335 }, 00:15:29.335 { 00:15:29.335 "name": "BaseBdev2", 00:15:29.335 "uuid": "0a632a98-aa29-5e4f-8d5c-732f5f7d1023", 00:15:29.335 "is_configured": true, 00:15:29.335 "data_offset": 2048, 00:15:29.335 "data_size": 63488 00:15:29.335 }, 00:15:29.335 { 00:15:29.335 "name": "BaseBdev3", 00:15:29.335 "uuid": "b3ec0800-b7f3-543b-baa8-c25a0e9115d1", 00:15:29.335 "is_configured": true, 00:15:29.335 "data_offset": 2048, 00:15:29.335 "data_size": 63488 00:15:29.335 }, 00:15:29.335 { 00:15:29.335 "name": "BaseBdev4", 00:15:29.335 "uuid": "315998c8-8084-57d1-bce6-0a4b3d0b594d", 00:15:29.335 "is_configured": true, 00:15:29.335 "data_offset": 2048, 00:15:29.335 "data_size": 63488 00:15:29.335 } 00:15:29.335 ] 00:15:29.335 }' 00:15:29.335 15:19:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:29.335 15:19:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:29.335 15:19:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:29.594 15:19:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:29.594 15:19:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:30.531 15:19:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:30.531 15:19:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:30.531 15:19:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:30.531 15:19:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:30.531 15:19:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:30.531 15:19:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:30.531 15:19:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:30.531 15:19:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:30.531 15:19:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:30.531 15:19:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:30.531 15:19:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:30.531 15:19:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:30.531 "name": "raid_bdev1", 00:15:30.531 "uuid": "e95708c2-d41d-45cb-9efa-c4b368f39949", 00:15:30.531 "strip_size_kb": 64, 00:15:30.531 "state": "online", 00:15:30.531 "raid_level": "raid5f", 00:15:30.531 "superblock": true, 00:15:30.531 "num_base_bdevs": 4, 00:15:30.531 "num_base_bdevs_discovered": 4, 00:15:30.531 "num_base_bdevs_operational": 4, 00:15:30.531 "process": { 00:15:30.531 "type": "rebuild", 00:15:30.531 "target": "spare", 00:15:30.531 "progress": { 00:15:30.531 "blocks": 153600, 00:15:30.531 "percent": 80 00:15:30.531 } 00:15:30.531 }, 00:15:30.531 "base_bdevs_list": [ 00:15:30.531 { 00:15:30.531 "name": "spare", 00:15:30.531 "uuid": "4f8f62cb-4a1f-59cd-a325-c384b3bd334f", 00:15:30.531 "is_configured": true, 00:15:30.531 "data_offset": 2048, 00:15:30.531 "data_size": 63488 00:15:30.531 }, 00:15:30.531 { 00:15:30.531 "name": "BaseBdev2", 00:15:30.531 "uuid": "0a632a98-aa29-5e4f-8d5c-732f5f7d1023", 00:15:30.531 "is_configured": true, 00:15:30.531 "data_offset": 2048, 00:15:30.531 "data_size": 63488 00:15:30.531 }, 00:15:30.531 { 00:15:30.531 "name": "BaseBdev3", 00:15:30.531 "uuid": "b3ec0800-b7f3-543b-baa8-c25a0e9115d1", 00:15:30.531 "is_configured": true, 00:15:30.531 "data_offset": 2048, 00:15:30.531 "data_size": 63488 00:15:30.531 }, 00:15:30.531 { 00:15:30.531 "name": "BaseBdev4", 00:15:30.531 "uuid": "315998c8-8084-57d1-bce6-0a4b3d0b594d", 00:15:30.531 "is_configured": true, 00:15:30.531 "data_offset": 2048, 00:15:30.531 "data_size": 63488 00:15:30.531 } 00:15:30.531 ] 00:15:30.531 }' 00:15:30.531 15:19:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:30.531 15:19:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:30.531 15:19:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:30.531 15:19:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:30.531 15:19:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:31.912 15:19:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:31.912 15:19:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:31.912 15:19:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:31.912 15:19:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:31.912 15:19:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:31.912 15:19:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:31.912 15:19:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:31.912 15:19:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:31.912 15:19:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:31.912 15:19:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:31.912 15:19:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:31.912 15:19:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:31.912 "name": "raid_bdev1", 00:15:31.912 "uuid": "e95708c2-d41d-45cb-9efa-c4b368f39949", 00:15:31.912 "strip_size_kb": 64, 00:15:31.912 "state": "online", 00:15:31.912 "raid_level": "raid5f", 00:15:31.912 "superblock": true, 00:15:31.912 "num_base_bdevs": 4, 00:15:31.912 "num_base_bdevs_discovered": 4, 00:15:31.912 "num_base_bdevs_operational": 4, 00:15:31.912 "process": { 00:15:31.912 "type": "rebuild", 00:15:31.912 "target": "spare", 00:15:31.912 "progress": { 00:15:31.912 "blocks": 176640, 00:15:31.912 "percent": 92 00:15:31.912 } 00:15:31.912 }, 00:15:31.912 "base_bdevs_list": [ 00:15:31.912 { 00:15:31.912 "name": "spare", 00:15:31.912 "uuid": "4f8f62cb-4a1f-59cd-a325-c384b3bd334f", 00:15:31.913 "is_configured": true, 00:15:31.913 "data_offset": 2048, 00:15:31.913 "data_size": 63488 00:15:31.913 }, 00:15:31.913 { 00:15:31.913 "name": "BaseBdev2", 00:15:31.913 "uuid": "0a632a98-aa29-5e4f-8d5c-732f5f7d1023", 00:15:31.913 "is_configured": true, 00:15:31.913 "data_offset": 2048, 00:15:31.913 "data_size": 63488 00:15:31.913 }, 00:15:31.913 { 00:15:31.913 "name": "BaseBdev3", 00:15:31.913 "uuid": "b3ec0800-b7f3-543b-baa8-c25a0e9115d1", 00:15:31.913 "is_configured": true, 00:15:31.913 "data_offset": 2048, 00:15:31.913 "data_size": 63488 00:15:31.913 }, 00:15:31.913 { 00:15:31.913 "name": "BaseBdev4", 00:15:31.913 "uuid": "315998c8-8084-57d1-bce6-0a4b3d0b594d", 00:15:31.913 "is_configured": true, 00:15:31.913 "data_offset": 2048, 00:15:31.913 "data_size": 63488 00:15:31.913 } 00:15:31.913 ] 00:15:31.913 }' 00:15:31.913 15:19:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:31.913 15:19:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:31.913 15:19:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:31.913 15:19:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:31.913 15:19:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:32.480 [2024-11-27 15:20:00.432838] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:15:32.480 [2024-11-27 15:20:00.432974] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:15:32.480 [2024-11-27 15:20:00.433128] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:32.738 15:20:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:32.738 15:20:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:32.738 15:20:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:32.738 15:20:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:32.738 15:20:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:32.738 15:20:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:32.738 15:20:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:32.738 15:20:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:32.738 15:20:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:32.738 15:20:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:32.738 15:20:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:32.738 15:20:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:32.738 "name": "raid_bdev1", 00:15:32.738 "uuid": "e95708c2-d41d-45cb-9efa-c4b368f39949", 00:15:32.738 "strip_size_kb": 64, 00:15:32.738 "state": "online", 00:15:32.738 "raid_level": "raid5f", 00:15:32.738 "superblock": true, 00:15:32.738 "num_base_bdevs": 4, 00:15:32.738 "num_base_bdevs_discovered": 4, 00:15:32.738 "num_base_bdevs_operational": 4, 00:15:32.738 "base_bdevs_list": [ 00:15:32.738 { 00:15:32.738 "name": "spare", 00:15:32.738 "uuid": "4f8f62cb-4a1f-59cd-a325-c384b3bd334f", 00:15:32.738 "is_configured": true, 00:15:32.738 "data_offset": 2048, 00:15:32.738 "data_size": 63488 00:15:32.738 }, 00:15:32.738 { 00:15:32.738 "name": "BaseBdev2", 00:15:32.738 "uuid": "0a632a98-aa29-5e4f-8d5c-732f5f7d1023", 00:15:32.738 "is_configured": true, 00:15:32.738 "data_offset": 2048, 00:15:32.738 "data_size": 63488 00:15:32.738 }, 00:15:32.738 { 00:15:32.738 "name": "BaseBdev3", 00:15:32.738 "uuid": "b3ec0800-b7f3-543b-baa8-c25a0e9115d1", 00:15:32.738 "is_configured": true, 00:15:32.738 "data_offset": 2048, 00:15:32.738 "data_size": 63488 00:15:32.738 }, 00:15:32.738 { 00:15:32.738 "name": "BaseBdev4", 00:15:32.738 "uuid": "315998c8-8084-57d1-bce6-0a4b3d0b594d", 00:15:32.738 "is_configured": true, 00:15:32.738 "data_offset": 2048, 00:15:32.738 "data_size": 63488 00:15:32.738 } 00:15:32.738 ] 00:15:32.738 }' 00:15:32.738 15:20:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:32.998 15:20:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:15:32.998 15:20:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:32.998 15:20:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:15:32.998 15:20:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:15:32.998 15:20:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:32.998 15:20:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:32.998 15:20:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:32.998 15:20:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:32.998 15:20:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:32.998 15:20:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:32.998 15:20:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:32.998 15:20:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:32.998 15:20:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:32.998 15:20:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:32.998 15:20:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:32.998 "name": "raid_bdev1", 00:15:32.998 "uuid": "e95708c2-d41d-45cb-9efa-c4b368f39949", 00:15:32.998 "strip_size_kb": 64, 00:15:32.998 "state": "online", 00:15:32.998 "raid_level": "raid5f", 00:15:32.998 "superblock": true, 00:15:32.998 "num_base_bdevs": 4, 00:15:32.998 "num_base_bdevs_discovered": 4, 00:15:32.998 "num_base_bdevs_operational": 4, 00:15:32.998 "base_bdevs_list": [ 00:15:32.998 { 00:15:32.998 "name": "spare", 00:15:32.998 "uuid": "4f8f62cb-4a1f-59cd-a325-c384b3bd334f", 00:15:32.998 "is_configured": true, 00:15:32.998 "data_offset": 2048, 00:15:32.998 "data_size": 63488 00:15:32.998 }, 00:15:32.998 { 00:15:32.998 "name": "BaseBdev2", 00:15:32.998 "uuid": "0a632a98-aa29-5e4f-8d5c-732f5f7d1023", 00:15:32.998 "is_configured": true, 00:15:32.998 "data_offset": 2048, 00:15:32.998 "data_size": 63488 00:15:32.998 }, 00:15:32.998 { 00:15:32.998 "name": "BaseBdev3", 00:15:32.998 "uuid": "b3ec0800-b7f3-543b-baa8-c25a0e9115d1", 00:15:32.998 "is_configured": true, 00:15:32.998 "data_offset": 2048, 00:15:32.998 "data_size": 63488 00:15:32.998 }, 00:15:32.998 { 00:15:32.998 "name": "BaseBdev4", 00:15:32.998 "uuid": "315998c8-8084-57d1-bce6-0a4b3d0b594d", 00:15:32.998 "is_configured": true, 00:15:32.998 "data_offset": 2048, 00:15:32.998 "data_size": 63488 00:15:32.998 } 00:15:32.998 ] 00:15:32.998 }' 00:15:32.998 15:20:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:32.998 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:32.998 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:32.998 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:32.998 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:15:32.998 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:32.998 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:32.998 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:32.998 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:32.998 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:32.998 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:32.998 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:32.998 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:32.998 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:32.998 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:32.998 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:32.998 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:32.998 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:32.998 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:33.257 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:33.257 "name": "raid_bdev1", 00:15:33.257 "uuid": "e95708c2-d41d-45cb-9efa-c4b368f39949", 00:15:33.257 "strip_size_kb": 64, 00:15:33.257 "state": "online", 00:15:33.257 "raid_level": "raid5f", 00:15:33.257 "superblock": true, 00:15:33.257 "num_base_bdevs": 4, 00:15:33.257 "num_base_bdevs_discovered": 4, 00:15:33.257 "num_base_bdevs_operational": 4, 00:15:33.257 "base_bdevs_list": [ 00:15:33.257 { 00:15:33.257 "name": "spare", 00:15:33.257 "uuid": "4f8f62cb-4a1f-59cd-a325-c384b3bd334f", 00:15:33.257 "is_configured": true, 00:15:33.257 "data_offset": 2048, 00:15:33.257 "data_size": 63488 00:15:33.257 }, 00:15:33.257 { 00:15:33.257 "name": "BaseBdev2", 00:15:33.257 "uuid": "0a632a98-aa29-5e4f-8d5c-732f5f7d1023", 00:15:33.258 "is_configured": true, 00:15:33.258 "data_offset": 2048, 00:15:33.258 "data_size": 63488 00:15:33.258 }, 00:15:33.258 { 00:15:33.258 "name": "BaseBdev3", 00:15:33.258 "uuid": "b3ec0800-b7f3-543b-baa8-c25a0e9115d1", 00:15:33.258 "is_configured": true, 00:15:33.258 "data_offset": 2048, 00:15:33.258 "data_size": 63488 00:15:33.258 }, 00:15:33.258 { 00:15:33.258 "name": "BaseBdev4", 00:15:33.258 "uuid": "315998c8-8084-57d1-bce6-0a4b3d0b594d", 00:15:33.258 "is_configured": true, 00:15:33.258 "data_offset": 2048, 00:15:33.258 "data_size": 63488 00:15:33.258 } 00:15:33.258 ] 00:15:33.258 }' 00:15:33.258 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:33.258 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:33.516 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:33.516 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:33.517 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:33.517 [2024-11-27 15:20:01.516997] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:33.517 [2024-11-27 15:20:01.517033] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:33.517 [2024-11-27 15:20:01.517123] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:33.517 [2024-11-27 15:20:01.517220] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:33.517 [2024-11-27 15:20:01.517243] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:15:33.517 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:33.517 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:33.517 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:15:33.517 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:33.517 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:33.517 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:33.517 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:15:33.517 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:15:33.517 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:15:33.517 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:15:33.517 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:33.517 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:15:33.517 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:33.517 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:33.517 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:33.517 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:15:33.517 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:33.517 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:33.517 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:15:33.776 /dev/nbd0 00:15:33.776 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:33.776 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:33.776 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:15:33.776 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:15:33.776 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:33.776 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:33.776 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:15:33.776 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:15:33.776 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:33.776 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:33.776 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:33.776 1+0 records in 00:15:33.776 1+0 records out 00:15:33.776 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000385244 s, 10.6 MB/s 00:15:33.776 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:33.776 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:15:33.776 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:33.776 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:33.776 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:15:33.776 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:33.776 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:33.776 15:20:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:15:34.036 /dev/nbd1 00:15:34.036 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:15:34.036 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:15:34.036 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:15:34.036 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:15:34.036 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:34.036 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:34.036 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:15:34.036 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:15:34.036 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:34.036 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:34.036 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:34.036 1+0 records in 00:15:34.036 1+0 records out 00:15:34.036 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000249043 s, 16.4 MB/s 00:15:34.036 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:34.036 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:15:34.036 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:34.036 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:34.036 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:15:34.036 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:34.036 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:34.036 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:15:34.036 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:15:34.036 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:34.036 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:34.036 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:34.036 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:15:34.036 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:34.036 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:34.295 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:34.295 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:34.295 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:34.295 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:34.295 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:34.295 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:34.295 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:15:34.295 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:15:34.295 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:34.295 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:15:34.556 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:15:34.556 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:15:34.556 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:15:34.556 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:34.556 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:34.556 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:15:34.556 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:15:34.556 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:15:34.556 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:15:34.556 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:15:34.556 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:34.556 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:34.556 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:34.556 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:34.556 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:34.556 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:34.556 [2024-11-27 15:20:02.551098] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:34.556 [2024-11-27 15:20:02.551152] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:34.556 [2024-11-27 15:20:02.551175] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:15:34.556 [2024-11-27 15:20:02.551185] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:34.556 [2024-11-27 15:20:02.553439] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:34.556 [2024-11-27 15:20:02.553483] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:34.556 [2024-11-27 15:20:02.553567] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:34.556 [2024-11-27 15:20:02.553614] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:34.556 [2024-11-27 15:20:02.553747] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:34.556 [2024-11-27 15:20:02.553857] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:34.556 [2024-11-27 15:20:02.553943] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:15:34.556 spare 00:15:34.556 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:34.556 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:15:34.556 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:34.556 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:34.556 [2024-11-27 15:20:02.653856] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006600 00:15:34.556 [2024-11-27 15:20:02.653883] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:34.556 [2024-11-27 15:20:02.654147] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000049030 00:15:34.556 [2024-11-27 15:20:02.654619] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006600 00:15:34.556 [2024-11-27 15:20:02.654643] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006600 00:15:34.556 [2024-11-27 15:20:02.654813] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:34.556 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:34.556 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:15:34.556 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:34.556 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:34.556 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:34.556 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:34.556 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:34.556 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:34.556 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:34.556 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:34.556 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:34.816 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:34.816 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:34.816 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:34.816 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:34.816 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:34.816 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:34.816 "name": "raid_bdev1", 00:15:34.816 "uuid": "e95708c2-d41d-45cb-9efa-c4b368f39949", 00:15:34.816 "strip_size_kb": 64, 00:15:34.816 "state": "online", 00:15:34.816 "raid_level": "raid5f", 00:15:34.816 "superblock": true, 00:15:34.816 "num_base_bdevs": 4, 00:15:34.816 "num_base_bdevs_discovered": 4, 00:15:34.816 "num_base_bdevs_operational": 4, 00:15:34.816 "base_bdevs_list": [ 00:15:34.816 { 00:15:34.816 "name": "spare", 00:15:34.816 "uuid": "4f8f62cb-4a1f-59cd-a325-c384b3bd334f", 00:15:34.816 "is_configured": true, 00:15:34.816 "data_offset": 2048, 00:15:34.816 "data_size": 63488 00:15:34.816 }, 00:15:34.816 { 00:15:34.816 "name": "BaseBdev2", 00:15:34.816 "uuid": "0a632a98-aa29-5e4f-8d5c-732f5f7d1023", 00:15:34.816 "is_configured": true, 00:15:34.816 "data_offset": 2048, 00:15:34.816 "data_size": 63488 00:15:34.816 }, 00:15:34.816 { 00:15:34.816 "name": "BaseBdev3", 00:15:34.816 "uuid": "b3ec0800-b7f3-543b-baa8-c25a0e9115d1", 00:15:34.816 "is_configured": true, 00:15:34.817 "data_offset": 2048, 00:15:34.817 "data_size": 63488 00:15:34.817 }, 00:15:34.817 { 00:15:34.817 "name": "BaseBdev4", 00:15:34.817 "uuid": "315998c8-8084-57d1-bce6-0a4b3d0b594d", 00:15:34.817 "is_configured": true, 00:15:34.817 "data_offset": 2048, 00:15:34.817 "data_size": 63488 00:15:34.817 } 00:15:34.817 ] 00:15:34.817 }' 00:15:34.817 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:34.817 15:20:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:35.076 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:35.076 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:35.076 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:35.076 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:35.076 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:35.076 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:35.076 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:35.076 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:35.076 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:35.076 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:35.076 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:35.076 "name": "raid_bdev1", 00:15:35.076 "uuid": "e95708c2-d41d-45cb-9efa-c4b368f39949", 00:15:35.076 "strip_size_kb": 64, 00:15:35.076 "state": "online", 00:15:35.076 "raid_level": "raid5f", 00:15:35.076 "superblock": true, 00:15:35.076 "num_base_bdevs": 4, 00:15:35.076 "num_base_bdevs_discovered": 4, 00:15:35.076 "num_base_bdevs_operational": 4, 00:15:35.076 "base_bdevs_list": [ 00:15:35.076 { 00:15:35.076 "name": "spare", 00:15:35.076 "uuid": "4f8f62cb-4a1f-59cd-a325-c384b3bd334f", 00:15:35.076 "is_configured": true, 00:15:35.076 "data_offset": 2048, 00:15:35.076 "data_size": 63488 00:15:35.076 }, 00:15:35.076 { 00:15:35.076 "name": "BaseBdev2", 00:15:35.076 "uuid": "0a632a98-aa29-5e4f-8d5c-732f5f7d1023", 00:15:35.076 "is_configured": true, 00:15:35.076 "data_offset": 2048, 00:15:35.076 "data_size": 63488 00:15:35.076 }, 00:15:35.076 { 00:15:35.076 "name": "BaseBdev3", 00:15:35.076 "uuid": "b3ec0800-b7f3-543b-baa8-c25a0e9115d1", 00:15:35.076 "is_configured": true, 00:15:35.076 "data_offset": 2048, 00:15:35.076 "data_size": 63488 00:15:35.076 }, 00:15:35.076 { 00:15:35.076 "name": "BaseBdev4", 00:15:35.076 "uuid": "315998c8-8084-57d1-bce6-0a4b3d0b594d", 00:15:35.076 "is_configured": true, 00:15:35.076 "data_offset": 2048, 00:15:35.076 "data_size": 63488 00:15:35.076 } 00:15:35.076 ] 00:15:35.076 }' 00:15:35.076 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:35.336 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:35.336 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:35.336 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:35.336 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:35.336 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:15:35.336 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:35.336 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:35.336 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:35.336 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:15:35.336 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:35.336 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:35.336 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:35.336 [2024-11-27 15:20:03.329851] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:35.336 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:35.336 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:35.336 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:35.336 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:35.336 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:35.336 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:35.336 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:35.336 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:35.336 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:35.336 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:35.336 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:35.336 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:35.336 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:35.336 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:35.336 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:35.336 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:35.336 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:35.336 "name": "raid_bdev1", 00:15:35.336 "uuid": "e95708c2-d41d-45cb-9efa-c4b368f39949", 00:15:35.336 "strip_size_kb": 64, 00:15:35.336 "state": "online", 00:15:35.336 "raid_level": "raid5f", 00:15:35.336 "superblock": true, 00:15:35.336 "num_base_bdevs": 4, 00:15:35.336 "num_base_bdevs_discovered": 3, 00:15:35.336 "num_base_bdevs_operational": 3, 00:15:35.336 "base_bdevs_list": [ 00:15:35.336 { 00:15:35.336 "name": null, 00:15:35.336 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:35.336 "is_configured": false, 00:15:35.336 "data_offset": 0, 00:15:35.336 "data_size": 63488 00:15:35.336 }, 00:15:35.336 { 00:15:35.336 "name": "BaseBdev2", 00:15:35.336 "uuid": "0a632a98-aa29-5e4f-8d5c-732f5f7d1023", 00:15:35.336 "is_configured": true, 00:15:35.336 "data_offset": 2048, 00:15:35.336 "data_size": 63488 00:15:35.336 }, 00:15:35.336 { 00:15:35.336 "name": "BaseBdev3", 00:15:35.336 "uuid": "b3ec0800-b7f3-543b-baa8-c25a0e9115d1", 00:15:35.336 "is_configured": true, 00:15:35.336 "data_offset": 2048, 00:15:35.336 "data_size": 63488 00:15:35.336 }, 00:15:35.336 { 00:15:35.336 "name": "BaseBdev4", 00:15:35.336 "uuid": "315998c8-8084-57d1-bce6-0a4b3d0b594d", 00:15:35.336 "is_configured": true, 00:15:35.336 "data_offset": 2048, 00:15:35.336 "data_size": 63488 00:15:35.336 } 00:15:35.336 ] 00:15:35.336 }' 00:15:35.336 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:35.336 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:35.906 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:35.906 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:35.906 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:35.906 [2024-11-27 15:20:03.789076] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:35.906 [2024-11-27 15:20:03.789283] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:35.906 [2024-11-27 15:20:03.789310] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:35.906 [2024-11-27 15:20:03.789348] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:35.906 [2024-11-27 15:20:03.793613] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000049100 00:15:35.906 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:35.906 15:20:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:15:35.906 [2024-11-27 15:20:03.795839] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:36.846 15:20:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:36.846 15:20:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:36.846 15:20:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:36.846 15:20:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:36.846 15:20:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:36.846 15:20:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:36.846 15:20:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:36.846 15:20:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.846 15:20:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:36.846 15:20:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.846 15:20:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:36.846 "name": "raid_bdev1", 00:15:36.846 "uuid": "e95708c2-d41d-45cb-9efa-c4b368f39949", 00:15:36.846 "strip_size_kb": 64, 00:15:36.846 "state": "online", 00:15:36.846 "raid_level": "raid5f", 00:15:36.846 "superblock": true, 00:15:36.846 "num_base_bdevs": 4, 00:15:36.846 "num_base_bdevs_discovered": 4, 00:15:36.846 "num_base_bdevs_operational": 4, 00:15:36.846 "process": { 00:15:36.846 "type": "rebuild", 00:15:36.846 "target": "spare", 00:15:36.846 "progress": { 00:15:36.846 "blocks": 19200, 00:15:36.846 "percent": 10 00:15:36.846 } 00:15:36.846 }, 00:15:36.846 "base_bdevs_list": [ 00:15:36.846 { 00:15:36.846 "name": "spare", 00:15:36.846 "uuid": "4f8f62cb-4a1f-59cd-a325-c384b3bd334f", 00:15:36.846 "is_configured": true, 00:15:36.846 "data_offset": 2048, 00:15:36.846 "data_size": 63488 00:15:36.846 }, 00:15:36.846 { 00:15:36.846 "name": "BaseBdev2", 00:15:36.846 "uuid": "0a632a98-aa29-5e4f-8d5c-732f5f7d1023", 00:15:36.846 "is_configured": true, 00:15:36.846 "data_offset": 2048, 00:15:36.846 "data_size": 63488 00:15:36.846 }, 00:15:36.846 { 00:15:36.846 "name": "BaseBdev3", 00:15:36.846 "uuid": "b3ec0800-b7f3-543b-baa8-c25a0e9115d1", 00:15:36.846 "is_configured": true, 00:15:36.846 "data_offset": 2048, 00:15:36.846 "data_size": 63488 00:15:36.846 }, 00:15:36.846 { 00:15:36.846 "name": "BaseBdev4", 00:15:36.846 "uuid": "315998c8-8084-57d1-bce6-0a4b3d0b594d", 00:15:36.846 "is_configured": true, 00:15:36.846 "data_offset": 2048, 00:15:36.846 "data_size": 63488 00:15:36.846 } 00:15:36.846 ] 00:15:36.846 }' 00:15:36.846 15:20:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:36.846 15:20:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:36.846 15:20:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:36.846 15:20:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:36.846 15:20:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:15:36.847 15:20:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.847 15:20:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:36.847 [2024-11-27 15:20:04.936383] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:37.107 [2024-11-27 15:20:05.002377] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:37.107 [2024-11-27 15:20:05.002451] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:37.107 [2024-11-27 15:20:05.002470] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:37.107 [2024-11-27 15:20:05.002478] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:37.107 15:20:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:37.107 15:20:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:37.107 15:20:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:37.107 15:20:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:37.107 15:20:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:37.107 15:20:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:37.107 15:20:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:37.107 15:20:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:37.107 15:20:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:37.107 15:20:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:37.107 15:20:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:37.107 15:20:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:37.107 15:20:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:37.107 15:20:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:37.107 15:20:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:37.107 15:20:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:37.107 15:20:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:37.107 "name": "raid_bdev1", 00:15:37.107 "uuid": "e95708c2-d41d-45cb-9efa-c4b368f39949", 00:15:37.107 "strip_size_kb": 64, 00:15:37.107 "state": "online", 00:15:37.107 "raid_level": "raid5f", 00:15:37.107 "superblock": true, 00:15:37.107 "num_base_bdevs": 4, 00:15:37.107 "num_base_bdevs_discovered": 3, 00:15:37.107 "num_base_bdevs_operational": 3, 00:15:37.107 "base_bdevs_list": [ 00:15:37.107 { 00:15:37.107 "name": null, 00:15:37.107 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:37.107 "is_configured": false, 00:15:37.107 "data_offset": 0, 00:15:37.107 "data_size": 63488 00:15:37.107 }, 00:15:37.107 { 00:15:37.107 "name": "BaseBdev2", 00:15:37.107 "uuid": "0a632a98-aa29-5e4f-8d5c-732f5f7d1023", 00:15:37.107 "is_configured": true, 00:15:37.107 "data_offset": 2048, 00:15:37.107 "data_size": 63488 00:15:37.107 }, 00:15:37.107 { 00:15:37.107 "name": "BaseBdev3", 00:15:37.107 "uuid": "b3ec0800-b7f3-543b-baa8-c25a0e9115d1", 00:15:37.107 "is_configured": true, 00:15:37.107 "data_offset": 2048, 00:15:37.107 "data_size": 63488 00:15:37.107 }, 00:15:37.107 { 00:15:37.107 "name": "BaseBdev4", 00:15:37.107 "uuid": "315998c8-8084-57d1-bce6-0a4b3d0b594d", 00:15:37.107 "is_configured": true, 00:15:37.107 "data_offset": 2048, 00:15:37.107 "data_size": 63488 00:15:37.107 } 00:15:37.107 ] 00:15:37.107 }' 00:15:37.107 15:20:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:37.107 15:20:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:37.368 15:20:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:37.368 15:20:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:37.368 15:20:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:37.368 [2024-11-27 15:20:05.455133] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:37.368 [2024-11-27 15:20:05.455241] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:37.368 [2024-11-27 15:20:05.455291] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c380 00:15:37.368 [2024-11-27 15:20:05.455320] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:37.368 [2024-11-27 15:20:05.455822] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:37.368 [2024-11-27 15:20:05.455892] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:37.368 [2024-11-27 15:20:05.456027] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:37.368 [2024-11-27 15:20:05.456071] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:37.368 [2024-11-27 15:20:05.456118] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:37.368 [2024-11-27 15:20:05.456183] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:37.368 [2024-11-27 15:20:05.460330] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000491d0 00:15:37.368 spare 00:15:37.368 15:20:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:37.368 15:20:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:15:37.368 [2024-11-27 15:20:05.462641] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:38.753 "name": "raid_bdev1", 00:15:38.753 "uuid": "e95708c2-d41d-45cb-9efa-c4b368f39949", 00:15:38.753 "strip_size_kb": 64, 00:15:38.753 "state": "online", 00:15:38.753 "raid_level": "raid5f", 00:15:38.753 "superblock": true, 00:15:38.753 "num_base_bdevs": 4, 00:15:38.753 "num_base_bdevs_discovered": 4, 00:15:38.753 "num_base_bdevs_operational": 4, 00:15:38.753 "process": { 00:15:38.753 "type": "rebuild", 00:15:38.753 "target": "spare", 00:15:38.753 "progress": { 00:15:38.753 "blocks": 19200, 00:15:38.753 "percent": 10 00:15:38.753 } 00:15:38.753 }, 00:15:38.753 "base_bdevs_list": [ 00:15:38.753 { 00:15:38.753 "name": "spare", 00:15:38.753 "uuid": "4f8f62cb-4a1f-59cd-a325-c384b3bd334f", 00:15:38.753 "is_configured": true, 00:15:38.753 "data_offset": 2048, 00:15:38.753 "data_size": 63488 00:15:38.753 }, 00:15:38.753 { 00:15:38.753 "name": "BaseBdev2", 00:15:38.753 "uuid": "0a632a98-aa29-5e4f-8d5c-732f5f7d1023", 00:15:38.753 "is_configured": true, 00:15:38.753 "data_offset": 2048, 00:15:38.753 "data_size": 63488 00:15:38.753 }, 00:15:38.753 { 00:15:38.753 "name": "BaseBdev3", 00:15:38.753 "uuid": "b3ec0800-b7f3-543b-baa8-c25a0e9115d1", 00:15:38.753 "is_configured": true, 00:15:38.753 "data_offset": 2048, 00:15:38.753 "data_size": 63488 00:15:38.753 }, 00:15:38.753 { 00:15:38.753 "name": "BaseBdev4", 00:15:38.753 "uuid": "315998c8-8084-57d1-bce6-0a4b3d0b594d", 00:15:38.753 "is_configured": true, 00:15:38.753 "data_offset": 2048, 00:15:38.753 "data_size": 63488 00:15:38.753 } 00:15:38.753 ] 00:15:38.753 }' 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:38.753 [2024-11-27 15:20:06.594477] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:38.753 [2024-11-27 15:20:06.669336] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:38.753 [2024-11-27 15:20:06.669468] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:38.753 [2024-11-27 15:20:06.669508] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:38.753 [2024-11-27 15:20:06.669549] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:38.753 "name": "raid_bdev1", 00:15:38.753 "uuid": "e95708c2-d41d-45cb-9efa-c4b368f39949", 00:15:38.753 "strip_size_kb": 64, 00:15:38.753 "state": "online", 00:15:38.753 "raid_level": "raid5f", 00:15:38.753 "superblock": true, 00:15:38.753 "num_base_bdevs": 4, 00:15:38.753 "num_base_bdevs_discovered": 3, 00:15:38.753 "num_base_bdevs_operational": 3, 00:15:38.753 "base_bdevs_list": [ 00:15:38.753 { 00:15:38.753 "name": null, 00:15:38.753 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:38.753 "is_configured": false, 00:15:38.753 "data_offset": 0, 00:15:38.753 "data_size": 63488 00:15:38.753 }, 00:15:38.753 { 00:15:38.753 "name": "BaseBdev2", 00:15:38.753 "uuid": "0a632a98-aa29-5e4f-8d5c-732f5f7d1023", 00:15:38.753 "is_configured": true, 00:15:38.753 "data_offset": 2048, 00:15:38.753 "data_size": 63488 00:15:38.753 }, 00:15:38.753 { 00:15:38.753 "name": "BaseBdev3", 00:15:38.753 "uuid": "b3ec0800-b7f3-543b-baa8-c25a0e9115d1", 00:15:38.753 "is_configured": true, 00:15:38.753 "data_offset": 2048, 00:15:38.753 "data_size": 63488 00:15:38.753 }, 00:15:38.753 { 00:15:38.753 "name": "BaseBdev4", 00:15:38.753 "uuid": "315998c8-8084-57d1-bce6-0a4b3d0b594d", 00:15:38.753 "is_configured": true, 00:15:38.753 "data_offset": 2048, 00:15:38.753 "data_size": 63488 00:15:38.753 } 00:15:38.753 ] 00:15:38.753 }' 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:38.753 15:20:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:39.014 15:20:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:39.014 15:20:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:39.275 15:20:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:39.275 15:20:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:39.276 15:20:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:39.276 15:20:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:39.276 15:20:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:39.276 15:20:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:39.276 15:20:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:39.276 15:20:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:39.276 15:20:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:39.276 "name": "raid_bdev1", 00:15:39.276 "uuid": "e95708c2-d41d-45cb-9efa-c4b368f39949", 00:15:39.276 "strip_size_kb": 64, 00:15:39.276 "state": "online", 00:15:39.276 "raid_level": "raid5f", 00:15:39.276 "superblock": true, 00:15:39.276 "num_base_bdevs": 4, 00:15:39.276 "num_base_bdevs_discovered": 3, 00:15:39.276 "num_base_bdevs_operational": 3, 00:15:39.276 "base_bdevs_list": [ 00:15:39.276 { 00:15:39.276 "name": null, 00:15:39.276 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:39.276 "is_configured": false, 00:15:39.276 "data_offset": 0, 00:15:39.276 "data_size": 63488 00:15:39.276 }, 00:15:39.276 { 00:15:39.276 "name": "BaseBdev2", 00:15:39.276 "uuid": "0a632a98-aa29-5e4f-8d5c-732f5f7d1023", 00:15:39.276 "is_configured": true, 00:15:39.276 "data_offset": 2048, 00:15:39.276 "data_size": 63488 00:15:39.276 }, 00:15:39.276 { 00:15:39.276 "name": "BaseBdev3", 00:15:39.276 "uuid": "b3ec0800-b7f3-543b-baa8-c25a0e9115d1", 00:15:39.276 "is_configured": true, 00:15:39.276 "data_offset": 2048, 00:15:39.276 "data_size": 63488 00:15:39.276 }, 00:15:39.276 { 00:15:39.276 "name": "BaseBdev4", 00:15:39.276 "uuid": "315998c8-8084-57d1-bce6-0a4b3d0b594d", 00:15:39.276 "is_configured": true, 00:15:39.276 "data_offset": 2048, 00:15:39.276 "data_size": 63488 00:15:39.276 } 00:15:39.276 ] 00:15:39.276 }' 00:15:39.276 15:20:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:39.276 15:20:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:39.276 15:20:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:39.276 15:20:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:39.276 15:20:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:15:39.276 15:20:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:39.276 15:20:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:39.276 15:20:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:39.276 15:20:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:39.276 15:20:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:39.276 15:20:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:39.276 [2024-11-27 15:20:07.254068] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:39.276 [2024-11-27 15:20:07.254185] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:39.276 [2024-11-27 15:20:07.254223] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c980 00:15:39.276 [2024-11-27 15:20:07.254260] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:39.276 [2024-11-27 15:20:07.254720] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:39.276 [2024-11-27 15:20:07.254782] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:39.276 [2024-11-27 15:20:07.254879] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:15:39.276 [2024-11-27 15:20:07.254945] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:39.276 [2024-11-27 15:20:07.254985] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:39.276 [2024-11-27 15:20:07.255046] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:15:39.276 BaseBdev1 00:15:39.276 15:20:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:39.276 15:20:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:15:40.230 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:40.230 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:40.230 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:40.230 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:40.230 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:40.230 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:40.230 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:40.230 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:40.230 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:40.230 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:40.230 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:40.230 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:40.230 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.230 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:40.230 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:40.230 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:40.230 "name": "raid_bdev1", 00:15:40.230 "uuid": "e95708c2-d41d-45cb-9efa-c4b368f39949", 00:15:40.230 "strip_size_kb": 64, 00:15:40.230 "state": "online", 00:15:40.230 "raid_level": "raid5f", 00:15:40.230 "superblock": true, 00:15:40.230 "num_base_bdevs": 4, 00:15:40.230 "num_base_bdevs_discovered": 3, 00:15:40.230 "num_base_bdevs_operational": 3, 00:15:40.230 "base_bdevs_list": [ 00:15:40.230 { 00:15:40.230 "name": null, 00:15:40.230 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:40.230 "is_configured": false, 00:15:40.230 "data_offset": 0, 00:15:40.230 "data_size": 63488 00:15:40.230 }, 00:15:40.230 { 00:15:40.230 "name": "BaseBdev2", 00:15:40.230 "uuid": "0a632a98-aa29-5e4f-8d5c-732f5f7d1023", 00:15:40.230 "is_configured": true, 00:15:40.230 "data_offset": 2048, 00:15:40.230 "data_size": 63488 00:15:40.230 }, 00:15:40.230 { 00:15:40.230 "name": "BaseBdev3", 00:15:40.230 "uuid": "b3ec0800-b7f3-543b-baa8-c25a0e9115d1", 00:15:40.230 "is_configured": true, 00:15:40.230 "data_offset": 2048, 00:15:40.230 "data_size": 63488 00:15:40.230 }, 00:15:40.230 { 00:15:40.230 "name": "BaseBdev4", 00:15:40.230 "uuid": "315998c8-8084-57d1-bce6-0a4b3d0b594d", 00:15:40.230 "is_configured": true, 00:15:40.230 "data_offset": 2048, 00:15:40.230 "data_size": 63488 00:15:40.230 } 00:15:40.230 ] 00:15:40.230 }' 00:15:40.230 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:40.230 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:40.813 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:40.813 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:40.813 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:40.813 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:40.813 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:40.813 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:40.813 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.813 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:40.813 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:40.813 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:40.813 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:40.813 "name": "raid_bdev1", 00:15:40.813 "uuid": "e95708c2-d41d-45cb-9efa-c4b368f39949", 00:15:40.813 "strip_size_kb": 64, 00:15:40.813 "state": "online", 00:15:40.813 "raid_level": "raid5f", 00:15:40.813 "superblock": true, 00:15:40.813 "num_base_bdevs": 4, 00:15:40.813 "num_base_bdevs_discovered": 3, 00:15:40.813 "num_base_bdevs_operational": 3, 00:15:40.813 "base_bdevs_list": [ 00:15:40.813 { 00:15:40.813 "name": null, 00:15:40.813 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:40.813 "is_configured": false, 00:15:40.813 "data_offset": 0, 00:15:40.813 "data_size": 63488 00:15:40.813 }, 00:15:40.813 { 00:15:40.813 "name": "BaseBdev2", 00:15:40.813 "uuid": "0a632a98-aa29-5e4f-8d5c-732f5f7d1023", 00:15:40.813 "is_configured": true, 00:15:40.813 "data_offset": 2048, 00:15:40.813 "data_size": 63488 00:15:40.813 }, 00:15:40.813 { 00:15:40.813 "name": "BaseBdev3", 00:15:40.813 "uuid": "b3ec0800-b7f3-543b-baa8-c25a0e9115d1", 00:15:40.813 "is_configured": true, 00:15:40.813 "data_offset": 2048, 00:15:40.813 "data_size": 63488 00:15:40.813 }, 00:15:40.813 { 00:15:40.813 "name": "BaseBdev4", 00:15:40.813 "uuid": "315998c8-8084-57d1-bce6-0a4b3d0b594d", 00:15:40.813 "is_configured": true, 00:15:40.813 "data_offset": 2048, 00:15:40.813 "data_size": 63488 00:15:40.813 } 00:15:40.813 ] 00:15:40.813 }' 00:15:40.813 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:40.813 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:40.813 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:40.813 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:40.813 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:40.813 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@652 -- # local es=0 00:15:40.813 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:40.813 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:15:40.813 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:40.813 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:15:40.813 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:40.813 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:40.813 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.813 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:40.813 [2024-11-27 15:20:08.887517] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:40.813 [2024-11-27 15:20:08.887744] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:40.813 [2024-11-27 15:20:08.887810] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:40.813 request: 00:15:40.813 { 00:15:40.813 "base_bdev": "BaseBdev1", 00:15:40.813 "raid_bdev": "raid_bdev1", 00:15:40.813 "method": "bdev_raid_add_base_bdev", 00:15:40.813 "req_id": 1 00:15:40.813 } 00:15:40.813 Got JSON-RPC error response 00:15:40.813 response: 00:15:40.813 { 00:15:40.813 "code": -22, 00:15:40.813 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:15:40.813 } 00:15:40.813 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:15:40.813 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@655 -- # es=1 00:15:40.813 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:15:40.813 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:15:40.813 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:15:40.813 15:20:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:15:42.194 15:20:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:42.194 15:20:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:42.194 15:20:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:42.194 15:20:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:42.194 15:20:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:42.194 15:20:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:42.194 15:20:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:42.194 15:20:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:42.194 15:20:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:42.194 15:20:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:42.194 15:20:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:42.194 15:20:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:42.194 15:20:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.194 15:20:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:42.194 15:20:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.194 15:20:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:42.194 "name": "raid_bdev1", 00:15:42.194 "uuid": "e95708c2-d41d-45cb-9efa-c4b368f39949", 00:15:42.194 "strip_size_kb": 64, 00:15:42.194 "state": "online", 00:15:42.194 "raid_level": "raid5f", 00:15:42.194 "superblock": true, 00:15:42.194 "num_base_bdevs": 4, 00:15:42.194 "num_base_bdevs_discovered": 3, 00:15:42.194 "num_base_bdevs_operational": 3, 00:15:42.194 "base_bdevs_list": [ 00:15:42.194 { 00:15:42.194 "name": null, 00:15:42.194 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:42.194 "is_configured": false, 00:15:42.194 "data_offset": 0, 00:15:42.194 "data_size": 63488 00:15:42.194 }, 00:15:42.194 { 00:15:42.194 "name": "BaseBdev2", 00:15:42.194 "uuid": "0a632a98-aa29-5e4f-8d5c-732f5f7d1023", 00:15:42.194 "is_configured": true, 00:15:42.194 "data_offset": 2048, 00:15:42.194 "data_size": 63488 00:15:42.194 }, 00:15:42.194 { 00:15:42.194 "name": "BaseBdev3", 00:15:42.194 "uuid": "b3ec0800-b7f3-543b-baa8-c25a0e9115d1", 00:15:42.194 "is_configured": true, 00:15:42.194 "data_offset": 2048, 00:15:42.194 "data_size": 63488 00:15:42.194 }, 00:15:42.194 { 00:15:42.194 "name": "BaseBdev4", 00:15:42.194 "uuid": "315998c8-8084-57d1-bce6-0a4b3d0b594d", 00:15:42.194 "is_configured": true, 00:15:42.194 "data_offset": 2048, 00:15:42.194 "data_size": 63488 00:15:42.194 } 00:15:42.194 ] 00:15:42.194 }' 00:15:42.194 15:20:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:42.194 15:20:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:42.453 15:20:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:42.453 15:20:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:42.453 15:20:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:42.453 15:20:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:42.453 15:20:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:42.453 15:20:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:42.453 15:20:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:42.453 15:20:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.453 15:20:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:42.453 15:20:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.453 15:20:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:42.453 "name": "raid_bdev1", 00:15:42.453 "uuid": "e95708c2-d41d-45cb-9efa-c4b368f39949", 00:15:42.453 "strip_size_kb": 64, 00:15:42.453 "state": "online", 00:15:42.453 "raid_level": "raid5f", 00:15:42.453 "superblock": true, 00:15:42.453 "num_base_bdevs": 4, 00:15:42.453 "num_base_bdevs_discovered": 3, 00:15:42.453 "num_base_bdevs_operational": 3, 00:15:42.453 "base_bdevs_list": [ 00:15:42.453 { 00:15:42.453 "name": null, 00:15:42.453 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:42.453 "is_configured": false, 00:15:42.453 "data_offset": 0, 00:15:42.453 "data_size": 63488 00:15:42.453 }, 00:15:42.453 { 00:15:42.453 "name": "BaseBdev2", 00:15:42.453 "uuid": "0a632a98-aa29-5e4f-8d5c-732f5f7d1023", 00:15:42.453 "is_configured": true, 00:15:42.453 "data_offset": 2048, 00:15:42.453 "data_size": 63488 00:15:42.453 }, 00:15:42.453 { 00:15:42.454 "name": "BaseBdev3", 00:15:42.454 "uuid": "b3ec0800-b7f3-543b-baa8-c25a0e9115d1", 00:15:42.454 "is_configured": true, 00:15:42.454 "data_offset": 2048, 00:15:42.454 "data_size": 63488 00:15:42.454 }, 00:15:42.454 { 00:15:42.454 "name": "BaseBdev4", 00:15:42.454 "uuid": "315998c8-8084-57d1-bce6-0a4b3d0b594d", 00:15:42.454 "is_configured": true, 00:15:42.454 "data_offset": 2048, 00:15:42.454 "data_size": 63488 00:15:42.454 } 00:15:42.454 ] 00:15:42.454 }' 00:15:42.454 15:20:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:42.454 15:20:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:42.454 15:20:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:42.454 15:20:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:42.454 15:20:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 95720 00:15:42.454 15:20:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@954 -- # '[' -z 95720 ']' 00:15:42.454 15:20:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@958 -- # kill -0 95720 00:15:42.454 15:20:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@959 -- # uname 00:15:42.454 15:20:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:42.454 15:20:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 95720 00:15:42.454 15:20:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:42.454 15:20:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:42.454 killing process with pid 95720 00:15:42.454 15:20:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 95720' 00:15:42.454 15:20:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@973 -- # kill 95720 00:15:42.454 Received shutdown signal, test time was about 60.000000 seconds 00:15:42.454 00:15:42.454 Latency(us) 00:15:42.454 [2024-11-27T15:20:10.561Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:15:42.454 [2024-11-27T15:20:10.561Z] =================================================================================================================== 00:15:42.454 [2024-11-27T15:20:10.561Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:15:42.454 [2024-11-27 15:20:10.522407] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:42.454 [2024-11-27 15:20:10.522531] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:42.454 15:20:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@978 -- # wait 95720 00:15:42.454 [2024-11-27 15:20:10.522613] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:42.454 [2024-11-27 15:20:10.522622] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state offline 00:15:42.714 [2024-11-27 15:20:10.575284] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:42.714 15:20:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:15:42.714 00:15:42.714 real 0m25.245s 00:15:42.714 user 0m32.107s 00:15:42.714 sys 0m3.048s 00:15:42.714 15:20:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:42.714 15:20:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:42.714 ************************************ 00:15:42.714 END TEST raid5f_rebuild_test_sb 00:15:42.714 ************************************ 00:15:42.974 15:20:10 bdev_raid -- bdev/bdev_raid.sh@995 -- # base_blocklen=4096 00:15:42.974 15:20:10 bdev_raid -- bdev/bdev_raid.sh@997 -- # run_test raid_state_function_test_sb_4k raid_state_function_test raid1 2 true 00:15:42.974 15:20:10 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:15:42.974 15:20:10 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:42.974 15:20:10 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:42.974 ************************************ 00:15:42.974 START TEST raid_state_function_test_sb_4k 00:15:42.974 ************************************ 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 true 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # local strip_size 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@229 -- # raid_pid=96513 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 96513' 00:15:42.974 Process raid pid: 96513 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@231 -- # waitforlisten 96513 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@835 -- # '[' -z 96513 ']' 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:42.974 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:42.974 15:20:10 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:42.974 [2024-11-27 15:20:10.943080] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:15:42.974 [2024-11-27 15:20:10.943300] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:43.235 [2024-11-27 15:20:11.112367] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:43.235 [2024-11-27 15:20:11.139044] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:43.235 [2024-11-27 15:20:11.182288] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:43.235 [2024-11-27 15:20:11.182326] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:43.803 15:20:11 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:43.803 15:20:11 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@868 -- # return 0 00:15:43.803 15:20:11 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:43.803 15:20:11 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:43.803 15:20:11 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:43.803 [2024-11-27 15:20:11.785311] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:43.803 [2024-11-27 15:20:11.785429] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:43.803 [2024-11-27 15:20:11.785462] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:43.803 [2024-11-27 15:20:11.785484] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:43.803 15:20:11 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:43.803 15:20:11 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:43.803 15:20:11 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:43.803 15:20:11 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:43.803 15:20:11 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:43.803 15:20:11 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:43.803 15:20:11 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:43.803 15:20:11 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:43.803 15:20:11 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:43.803 15:20:11 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:43.803 15:20:11 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:43.803 15:20:11 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:43.803 15:20:11 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:43.803 15:20:11 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:43.803 15:20:11 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:43.803 15:20:11 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:43.803 15:20:11 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:43.803 "name": "Existed_Raid", 00:15:43.803 "uuid": "9e44b043-0ab2-430f-9224-c600c1800935", 00:15:43.803 "strip_size_kb": 0, 00:15:43.803 "state": "configuring", 00:15:43.803 "raid_level": "raid1", 00:15:43.803 "superblock": true, 00:15:43.803 "num_base_bdevs": 2, 00:15:43.803 "num_base_bdevs_discovered": 0, 00:15:43.803 "num_base_bdevs_operational": 2, 00:15:43.803 "base_bdevs_list": [ 00:15:43.803 { 00:15:43.803 "name": "BaseBdev1", 00:15:43.803 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:43.803 "is_configured": false, 00:15:43.803 "data_offset": 0, 00:15:43.803 "data_size": 0 00:15:43.803 }, 00:15:43.803 { 00:15:43.803 "name": "BaseBdev2", 00:15:43.803 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:43.803 "is_configured": false, 00:15:43.803 "data_offset": 0, 00:15:43.803 "data_size": 0 00:15:43.803 } 00:15:43.803 ] 00:15:43.803 }' 00:15:43.803 15:20:11 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:43.803 15:20:11 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:44.372 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:44.372 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.372 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:44.372 [2024-11-27 15:20:12.228481] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:44.372 [2024-11-27 15:20:12.228522] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:15:44.372 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.372 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:44.372 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.372 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:44.372 [2024-11-27 15:20:12.240445] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:44.372 [2024-11-27 15:20:12.240526] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:44.372 [2024-11-27 15:20:12.240553] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:44.372 [2024-11-27 15:20:12.240575] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:44.372 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.372 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev1 00:15:44.372 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.372 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:44.372 [2024-11-27 15:20:12.261535] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:44.372 BaseBdev1 00:15:44.372 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.372 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:15:44.372 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:15:44.372 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:44.372 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@905 -- # local i 00:15:44.372 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:44.372 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:44.372 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:44.372 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.372 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:44.372 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.372 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:15:44.372 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.372 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:44.372 [ 00:15:44.372 { 00:15:44.372 "name": "BaseBdev1", 00:15:44.372 "aliases": [ 00:15:44.372 "fa7363d7-a993-42cd-aac6-a4e508a435af" 00:15:44.372 ], 00:15:44.372 "product_name": "Malloc disk", 00:15:44.372 "block_size": 4096, 00:15:44.372 "num_blocks": 8192, 00:15:44.372 "uuid": "fa7363d7-a993-42cd-aac6-a4e508a435af", 00:15:44.372 "assigned_rate_limits": { 00:15:44.372 "rw_ios_per_sec": 0, 00:15:44.372 "rw_mbytes_per_sec": 0, 00:15:44.372 "r_mbytes_per_sec": 0, 00:15:44.372 "w_mbytes_per_sec": 0 00:15:44.372 }, 00:15:44.372 "claimed": true, 00:15:44.372 "claim_type": "exclusive_write", 00:15:44.372 "zoned": false, 00:15:44.372 "supported_io_types": { 00:15:44.372 "read": true, 00:15:44.372 "write": true, 00:15:44.372 "unmap": true, 00:15:44.372 "flush": true, 00:15:44.372 "reset": true, 00:15:44.372 "nvme_admin": false, 00:15:44.372 "nvme_io": false, 00:15:44.372 "nvme_io_md": false, 00:15:44.372 "write_zeroes": true, 00:15:44.372 "zcopy": true, 00:15:44.372 "get_zone_info": false, 00:15:44.372 "zone_management": false, 00:15:44.372 "zone_append": false, 00:15:44.372 "compare": false, 00:15:44.372 "compare_and_write": false, 00:15:44.372 "abort": true, 00:15:44.372 "seek_hole": false, 00:15:44.372 "seek_data": false, 00:15:44.372 "copy": true, 00:15:44.372 "nvme_iov_md": false 00:15:44.372 }, 00:15:44.372 "memory_domains": [ 00:15:44.372 { 00:15:44.372 "dma_device_id": "system", 00:15:44.372 "dma_device_type": 1 00:15:44.372 }, 00:15:44.372 { 00:15:44.372 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:44.372 "dma_device_type": 2 00:15:44.372 } 00:15:44.372 ], 00:15:44.372 "driver_specific": {} 00:15:44.372 } 00:15:44.372 ] 00:15:44.372 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.372 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@911 -- # return 0 00:15:44.372 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:44.372 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:44.372 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:44.372 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:44.372 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:44.373 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:44.373 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:44.373 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:44.373 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:44.373 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:44.373 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:44.373 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.373 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:44.373 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:44.373 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.373 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:44.373 "name": "Existed_Raid", 00:15:44.373 "uuid": "03fb7bc3-ae7a-49c6-89a2-88a35d3bd0f6", 00:15:44.373 "strip_size_kb": 0, 00:15:44.373 "state": "configuring", 00:15:44.373 "raid_level": "raid1", 00:15:44.373 "superblock": true, 00:15:44.373 "num_base_bdevs": 2, 00:15:44.373 "num_base_bdevs_discovered": 1, 00:15:44.373 "num_base_bdevs_operational": 2, 00:15:44.373 "base_bdevs_list": [ 00:15:44.373 { 00:15:44.373 "name": "BaseBdev1", 00:15:44.373 "uuid": "fa7363d7-a993-42cd-aac6-a4e508a435af", 00:15:44.373 "is_configured": true, 00:15:44.373 "data_offset": 256, 00:15:44.373 "data_size": 7936 00:15:44.373 }, 00:15:44.373 { 00:15:44.373 "name": "BaseBdev2", 00:15:44.373 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:44.373 "is_configured": false, 00:15:44.373 "data_offset": 0, 00:15:44.373 "data_size": 0 00:15:44.373 } 00:15:44.373 ] 00:15:44.373 }' 00:15:44.373 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:44.373 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:44.632 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:44.632 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.632 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:44.632 [2024-11-27 15:20:12.732753] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:44.632 [2024-11-27 15:20:12.732862] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:15:44.892 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.892 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:44.892 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.892 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:44.892 [2024-11-27 15:20:12.744756] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:44.892 [2024-11-27 15:20:12.746613] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:44.892 [2024-11-27 15:20:12.746703] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:44.892 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.892 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:15:44.892 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:44.892 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:44.892 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:44.892 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:44.892 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:44.892 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:44.892 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:44.892 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:44.892 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:44.892 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:44.892 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:44.892 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:44.892 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:44.892 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.892 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:44.892 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.892 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:44.892 "name": "Existed_Raid", 00:15:44.892 "uuid": "193b5f8e-f402-40bd-b2ee-443b2d50024a", 00:15:44.892 "strip_size_kb": 0, 00:15:44.892 "state": "configuring", 00:15:44.892 "raid_level": "raid1", 00:15:44.892 "superblock": true, 00:15:44.892 "num_base_bdevs": 2, 00:15:44.892 "num_base_bdevs_discovered": 1, 00:15:44.892 "num_base_bdevs_operational": 2, 00:15:44.892 "base_bdevs_list": [ 00:15:44.892 { 00:15:44.892 "name": "BaseBdev1", 00:15:44.892 "uuid": "fa7363d7-a993-42cd-aac6-a4e508a435af", 00:15:44.892 "is_configured": true, 00:15:44.892 "data_offset": 256, 00:15:44.892 "data_size": 7936 00:15:44.892 }, 00:15:44.892 { 00:15:44.892 "name": "BaseBdev2", 00:15:44.892 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:44.892 "is_configured": false, 00:15:44.892 "data_offset": 0, 00:15:44.892 "data_size": 0 00:15:44.892 } 00:15:44.892 ] 00:15:44.892 }' 00:15:44.892 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:44.892 15:20:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:45.152 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev2 00:15:45.152 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.152 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:45.152 [2024-11-27 15:20:13.175175] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:45.152 [2024-11-27 15:20:13.175494] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:15:45.152 [2024-11-27 15:20:13.175546] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:45.152 BaseBdev2 00:15:45.152 [2024-11-27 15:20:13.175858] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:15:45.152 [2024-11-27 15:20:13.176032] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:15:45.152 [2024-11-27 15:20:13.176052] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:15:45.152 [2024-11-27 15:20:13.176172] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:45.152 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.152 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:15:45.152 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:15:45.152 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:45.152 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@905 -- # local i 00:15:45.152 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:45.152 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:45.152 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:45.152 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.152 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:45.152 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.152 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:15:45.153 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.153 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:45.153 [ 00:15:45.153 { 00:15:45.153 "name": "BaseBdev2", 00:15:45.153 "aliases": [ 00:15:45.153 "e972fb61-388f-4962-8741-51e4ed3ca48f" 00:15:45.153 ], 00:15:45.153 "product_name": "Malloc disk", 00:15:45.153 "block_size": 4096, 00:15:45.153 "num_blocks": 8192, 00:15:45.153 "uuid": "e972fb61-388f-4962-8741-51e4ed3ca48f", 00:15:45.153 "assigned_rate_limits": { 00:15:45.153 "rw_ios_per_sec": 0, 00:15:45.153 "rw_mbytes_per_sec": 0, 00:15:45.153 "r_mbytes_per_sec": 0, 00:15:45.153 "w_mbytes_per_sec": 0 00:15:45.153 }, 00:15:45.153 "claimed": true, 00:15:45.153 "claim_type": "exclusive_write", 00:15:45.153 "zoned": false, 00:15:45.153 "supported_io_types": { 00:15:45.153 "read": true, 00:15:45.153 "write": true, 00:15:45.153 "unmap": true, 00:15:45.153 "flush": true, 00:15:45.153 "reset": true, 00:15:45.153 "nvme_admin": false, 00:15:45.153 "nvme_io": false, 00:15:45.153 "nvme_io_md": false, 00:15:45.153 "write_zeroes": true, 00:15:45.153 "zcopy": true, 00:15:45.153 "get_zone_info": false, 00:15:45.153 "zone_management": false, 00:15:45.153 "zone_append": false, 00:15:45.153 "compare": false, 00:15:45.153 "compare_and_write": false, 00:15:45.153 "abort": true, 00:15:45.153 "seek_hole": false, 00:15:45.153 "seek_data": false, 00:15:45.153 "copy": true, 00:15:45.153 "nvme_iov_md": false 00:15:45.153 }, 00:15:45.153 "memory_domains": [ 00:15:45.153 { 00:15:45.153 "dma_device_id": "system", 00:15:45.153 "dma_device_type": 1 00:15:45.153 }, 00:15:45.153 { 00:15:45.153 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:45.153 "dma_device_type": 2 00:15:45.153 } 00:15:45.153 ], 00:15:45.153 "driver_specific": {} 00:15:45.153 } 00:15:45.153 ] 00:15:45.153 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.153 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@911 -- # return 0 00:15:45.153 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:15:45.153 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:45.153 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:15:45.153 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:45.153 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:45.153 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:45.153 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:45.153 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:45.153 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:45.153 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:45.153 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:45.153 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:45.153 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:45.153 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:45.153 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.153 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:45.153 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.412 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:45.412 "name": "Existed_Raid", 00:15:45.412 "uuid": "193b5f8e-f402-40bd-b2ee-443b2d50024a", 00:15:45.412 "strip_size_kb": 0, 00:15:45.412 "state": "online", 00:15:45.412 "raid_level": "raid1", 00:15:45.412 "superblock": true, 00:15:45.412 "num_base_bdevs": 2, 00:15:45.412 "num_base_bdevs_discovered": 2, 00:15:45.412 "num_base_bdevs_operational": 2, 00:15:45.412 "base_bdevs_list": [ 00:15:45.412 { 00:15:45.412 "name": "BaseBdev1", 00:15:45.412 "uuid": "fa7363d7-a993-42cd-aac6-a4e508a435af", 00:15:45.412 "is_configured": true, 00:15:45.412 "data_offset": 256, 00:15:45.412 "data_size": 7936 00:15:45.412 }, 00:15:45.412 { 00:15:45.412 "name": "BaseBdev2", 00:15:45.412 "uuid": "e972fb61-388f-4962-8741-51e4ed3ca48f", 00:15:45.412 "is_configured": true, 00:15:45.412 "data_offset": 256, 00:15:45.412 "data_size": 7936 00:15:45.412 } 00:15:45.412 ] 00:15:45.412 }' 00:15:45.412 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:45.412 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:45.672 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:15:45.672 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:15:45.672 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:45.672 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:45.672 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@184 -- # local name 00:15:45.672 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:45.672 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:15:45.672 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.672 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:45.673 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:45.673 [2024-11-27 15:20:13.654806] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:45.673 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.673 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:45.673 "name": "Existed_Raid", 00:15:45.673 "aliases": [ 00:15:45.673 "193b5f8e-f402-40bd-b2ee-443b2d50024a" 00:15:45.673 ], 00:15:45.673 "product_name": "Raid Volume", 00:15:45.673 "block_size": 4096, 00:15:45.673 "num_blocks": 7936, 00:15:45.673 "uuid": "193b5f8e-f402-40bd-b2ee-443b2d50024a", 00:15:45.673 "assigned_rate_limits": { 00:15:45.673 "rw_ios_per_sec": 0, 00:15:45.673 "rw_mbytes_per_sec": 0, 00:15:45.673 "r_mbytes_per_sec": 0, 00:15:45.673 "w_mbytes_per_sec": 0 00:15:45.673 }, 00:15:45.673 "claimed": false, 00:15:45.673 "zoned": false, 00:15:45.673 "supported_io_types": { 00:15:45.673 "read": true, 00:15:45.673 "write": true, 00:15:45.673 "unmap": false, 00:15:45.673 "flush": false, 00:15:45.673 "reset": true, 00:15:45.673 "nvme_admin": false, 00:15:45.673 "nvme_io": false, 00:15:45.673 "nvme_io_md": false, 00:15:45.673 "write_zeroes": true, 00:15:45.673 "zcopy": false, 00:15:45.673 "get_zone_info": false, 00:15:45.673 "zone_management": false, 00:15:45.673 "zone_append": false, 00:15:45.673 "compare": false, 00:15:45.673 "compare_and_write": false, 00:15:45.673 "abort": false, 00:15:45.673 "seek_hole": false, 00:15:45.673 "seek_data": false, 00:15:45.673 "copy": false, 00:15:45.673 "nvme_iov_md": false 00:15:45.673 }, 00:15:45.673 "memory_domains": [ 00:15:45.673 { 00:15:45.673 "dma_device_id": "system", 00:15:45.673 "dma_device_type": 1 00:15:45.673 }, 00:15:45.673 { 00:15:45.673 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:45.673 "dma_device_type": 2 00:15:45.673 }, 00:15:45.673 { 00:15:45.673 "dma_device_id": "system", 00:15:45.673 "dma_device_type": 1 00:15:45.673 }, 00:15:45.673 { 00:15:45.673 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:45.673 "dma_device_type": 2 00:15:45.673 } 00:15:45.673 ], 00:15:45.673 "driver_specific": { 00:15:45.673 "raid": { 00:15:45.673 "uuid": "193b5f8e-f402-40bd-b2ee-443b2d50024a", 00:15:45.673 "strip_size_kb": 0, 00:15:45.673 "state": "online", 00:15:45.673 "raid_level": "raid1", 00:15:45.673 "superblock": true, 00:15:45.673 "num_base_bdevs": 2, 00:15:45.673 "num_base_bdevs_discovered": 2, 00:15:45.673 "num_base_bdevs_operational": 2, 00:15:45.673 "base_bdevs_list": [ 00:15:45.673 { 00:15:45.673 "name": "BaseBdev1", 00:15:45.673 "uuid": "fa7363d7-a993-42cd-aac6-a4e508a435af", 00:15:45.673 "is_configured": true, 00:15:45.673 "data_offset": 256, 00:15:45.673 "data_size": 7936 00:15:45.673 }, 00:15:45.673 { 00:15:45.673 "name": "BaseBdev2", 00:15:45.673 "uuid": "e972fb61-388f-4962-8741-51e4ed3ca48f", 00:15:45.673 "is_configured": true, 00:15:45.673 "data_offset": 256, 00:15:45.673 "data_size": 7936 00:15:45.673 } 00:15:45.673 ] 00:15:45.673 } 00:15:45.673 } 00:15:45.673 }' 00:15:45.673 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:45.673 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:15:45.673 BaseBdev2' 00:15:45.673 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:45.673 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:15:45.673 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:45.933 [2024-11-27 15:20:13.866113] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@260 -- # local expected_state 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@198 -- # case $1 in 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@199 -- # return 0 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:45.933 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:45.934 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:45.934 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.934 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:45.934 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:45.934 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.934 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:45.934 "name": "Existed_Raid", 00:15:45.934 "uuid": "193b5f8e-f402-40bd-b2ee-443b2d50024a", 00:15:45.934 "strip_size_kb": 0, 00:15:45.934 "state": "online", 00:15:45.934 "raid_level": "raid1", 00:15:45.934 "superblock": true, 00:15:45.934 "num_base_bdevs": 2, 00:15:45.934 "num_base_bdevs_discovered": 1, 00:15:45.934 "num_base_bdevs_operational": 1, 00:15:45.934 "base_bdevs_list": [ 00:15:45.934 { 00:15:45.934 "name": null, 00:15:45.934 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:45.934 "is_configured": false, 00:15:45.934 "data_offset": 0, 00:15:45.934 "data_size": 7936 00:15:45.934 }, 00:15:45.934 { 00:15:45.934 "name": "BaseBdev2", 00:15:45.934 "uuid": "e972fb61-388f-4962-8741-51e4ed3ca48f", 00:15:45.934 "is_configured": true, 00:15:45.934 "data_offset": 256, 00:15:45.934 "data_size": 7936 00:15:45.934 } 00:15:45.934 ] 00:15:45.934 }' 00:15:45.934 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:45.934 15:20:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:46.504 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:15:46.504 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:46.504 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:46.504 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:46.504 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:15:46.504 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:46.504 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:46.504 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:15:46.504 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:15:46.504 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:15:46.504 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:46.504 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:46.504 [2024-11-27 15:20:14.370082] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:15:46.505 [2024-11-27 15:20:14.370260] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:46.505 [2024-11-27 15:20:14.391095] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:46.505 [2024-11-27 15:20:14.391225] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:46.505 [2024-11-27 15:20:14.391281] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:15:46.505 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:46.505 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:15:46.505 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:46.505 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:46.505 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:15:46.505 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:46.505 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:46.505 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:46.505 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:15:46.505 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:15:46.505 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:15:46.505 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@326 -- # killprocess 96513 00:15:46.505 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@954 -- # '[' -z 96513 ']' 00:15:46.505 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@958 -- # kill -0 96513 00:15:46.505 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@959 -- # uname 00:15:46.505 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:46.505 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 96513 00:15:46.505 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:46.505 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:46.505 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@972 -- # echo 'killing process with pid 96513' 00:15:46.505 killing process with pid 96513 00:15:46.505 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@973 -- # kill 96513 00:15:46.505 [2024-11-27 15:20:14.488849] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:46.505 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@978 -- # wait 96513 00:15:46.505 [2024-11-27 15:20:14.490415] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:46.764 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@328 -- # return 0 00:15:46.764 ************************************ 00:15:46.764 END TEST raid_state_function_test_sb_4k 00:15:46.764 ************************************ 00:15:46.764 00:15:46.764 real 0m3.969s 00:15:46.764 user 0m6.150s 00:15:46.764 sys 0m0.826s 00:15:46.764 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:46.764 15:20:14 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:47.024 15:20:14 bdev_raid -- bdev/bdev_raid.sh@998 -- # run_test raid_superblock_test_4k raid_superblock_test raid1 2 00:15:47.024 15:20:14 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:15:47.024 15:20:14 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:47.024 15:20:14 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:47.024 ************************************ 00:15:47.024 START TEST raid_superblock_test_4k 00:15:47.024 ************************************ 00:15:47.024 15:20:14 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 2 00:15:47.024 15:20:14 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:15:47.024 15:20:14 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:15:47.024 15:20:14 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:15:47.024 15:20:14 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:15:47.024 15:20:14 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:15:47.024 15:20:14 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:15:47.024 15:20:14 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:15:47.024 15:20:14 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:15:47.024 15:20:14 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:15:47.024 15:20:14 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@399 -- # local strip_size 00:15:47.024 15:20:14 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:15:47.024 15:20:14 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:15:47.024 15:20:14 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:15:47.024 15:20:14 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:15:47.024 15:20:14 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:15:47.024 15:20:14 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@412 -- # raid_pid=96754 00:15:47.024 15:20:14 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:15:47.024 15:20:14 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@413 -- # waitforlisten 96754 00:15:47.024 15:20:14 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@835 -- # '[' -z 96754 ']' 00:15:47.024 15:20:14 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:47.024 15:20:14 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:47.024 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:47.024 15:20:14 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:47.024 15:20:14 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:47.024 15:20:14 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:47.024 [2024-11-27 15:20:14.984272] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:15:47.024 [2024-11-27 15:20:14.984836] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid96754 ] 00:15:47.284 [2024-11-27 15:20:15.156169] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:47.284 [2024-11-27 15:20:15.183620] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:47.284 [2024-11-27 15:20:15.227552] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:47.284 [2024-11-27 15:20:15.227593] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:47.853 15:20:15 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:47.853 15:20:15 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@868 -- # return 0 00:15:47.853 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:15:47.853 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:47.853 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:15:47.853 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:15:47.853 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:15:47.853 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:47.853 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc1 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:47.854 malloc1 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:47.854 [2024-11-27 15:20:15.850089] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:47.854 [2024-11-27 15:20:15.850258] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:47.854 [2024-11-27 15:20:15.850307] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:47.854 [2024-11-27 15:20:15.850351] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:47.854 [2024-11-27 15:20:15.852752] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:47.854 [2024-11-27 15:20:15.852842] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:47.854 pt1 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc2 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:47.854 malloc2 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:47.854 [2024-11-27 15:20:15.884540] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:47.854 [2024-11-27 15:20:15.884653] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:47.854 [2024-11-27 15:20:15.884692] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:15:47.854 [2024-11-27 15:20:15.884727] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:47.854 [2024-11-27 15:20:15.887013] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:47.854 [2024-11-27 15:20:15.887096] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:47.854 pt2 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:47.854 [2024-11-27 15:20:15.896563] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:47.854 [2024-11-27 15:20:15.898604] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:47.854 [2024-11-27 15:20:15.898803] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:15:47.854 [2024-11-27 15:20:15.898857] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:47.854 [2024-11-27 15:20:15.899178] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:15:47.854 [2024-11-27 15:20:15.899375] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:15:47.854 [2024-11-27 15:20:15.899434] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:15:47.854 [2024-11-27 15:20:15.899608] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:47.854 15:20:15 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.191 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:48.191 "name": "raid_bdev1", 00:15:48.191 "uuid": "af8052d6-b181-460d-9bf2-b55908f07945", 00:15:48.191 "strip_size_kb": 0, 00:15:48.191 "state": "online", 00:15:48.191 "raid_level": "raid1", 00:15:48.191 "superblock": true, 00:15:48.191 "num_base_bdevs": 2, 00:15:48.191 "num_base_bdevs_discovered": 2, 00:15:48.191 "num_base_bdevs_operational": 2, 00:15:48.191 "base_bdevs_list": [ 00:15:48.191 { 00:15:48.191 "name": "pt1", 00:15:48.191 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:48.191 "is_configured": true, 00:15:48.191 "data_offset": 256, 00:15:48.191 "data_size": 7936 00:15:48.191 }, 00:15:48.191 { 00:15:48.191 "name": "pt2", 00:15:48.191 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:48.191 "is_configured": true, 00:15:48.191 "data_offset": 256, 00:15:48.191 "data_size": 7936 00:15:48.191 } 00:15:48.191 ] 00:15:48.191 }' 00:15:48.191 15:20:15 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:48.191 15:20:15 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:48.451 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:15:48.451 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:15:48.451 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:48.451 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:48.451 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@184 -- # local name 00:15:48.451 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:48.451 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:48.451 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:48.451 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.451 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:48.451 [2024-11-27 15:20:16.332110] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:48.451 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.451 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:48.451 "name": "raid_bdev1", 00:15:48.451 "aliases": [ 00:15:48.451 "af8052d6-b181-460d-9bf2-b55908f07945" 00:15:48.451 ], 00:15:48.451 "product_name": "Raid Volume", 00:15:48.451 "block_size": 4096, 00:15:48.451 "num_blocks": 7936, 00:15:48.451 "uuid": "af8052d6-b181-460d-9bf2-b55908f07945", 00:15:48.451 "assigned_rate_limits": { 00:15:48.451 "rw_ios_per_sec": 0, 00:15:48.451 "rw_mbytes_per_sec": 0, 00:15:48.451 "r_mbytes_per_sec": 0, 00:15:48.451 "w_mbytes_per_sec": 0 00:15:48.451 }, 00:15:48.451 "claimed": false, 00:15:48.451 "zoned": false, 00:15:48.451 "supported_io_types": { 00:15:48.451 "read": true, 00:15:48.451 "write": true, 00:15:48.451 "unmap": false, 00:15:48.451 "flush": false, 00:15:48.451 "reset": true, 00:15:48.451 "nvme_admin": false, 00:15:48.451 "nvme_io": false, 00:15:48.451 "nvme_io_md": false, 00:15:48.451 "write_zeroes": true, 00:15:48.451 "zcopy": false, 00:15:48.451 "get_zone_info": false, 00:15:48.451 "zone_management": false, 00:15:48.451 "zone_append": false, 00:15:48.451 "compare": false, 00:15:48.451 "compare_and_write": false, 00:15:48.451 "abort": false, 00:15:48.451 "seek_hole": false, 00:15:48.451 "seek_data": false, 00:15:48.451 "copy": false, 00:15:48.451 "nvme_iov_md": false 00:15:48.451 }, 00:15:48.451 "memory_domains": [ 00:15:48.451 { 00:15:48.451 "dma_device_id": "system", 00:15:48.451 "dma_device_type": 1 00:15:48.451 }, 00:15:48.451 { 00:15:48.451 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:48.451 "dma_device_type": 2 00:15:48.451 }, 00:15:48.451 { 00:15:48.451 "dma_device_id": "system", 00:15:48.451 "dma_device_type": 1 00:15:48.451 }, 00:15:48.451 { 00:15:48.451 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:48.451 "dma_device_type": 2 00:15:48.451 } 00:15:48.451 ], 00:15:48.451 "driver_specific": { 00:15:48.451 "raid": { 00:15:48.451 "uuid": "af8052d6-b181-460d-9bf2-b55908f07945", 00:15:48.451 "strip_size_kb": 0, 00:15:48.451 "state": "online", 00:15:48.451 "raid_level": "raid1", 00:15:48.451 "superblock": true, 00:15:48.451 "num_base_bdevs": 2, 00:15:48.451 "num_base_bdevs_discovered": 2, 00:15:48.451 "num_base_bdevs_operational": 2, 00:15:48.451 "base_bdevs_list": [ 00:15:48.451 { 00:15:48.451 "name": "pt1", 00:15:48.451 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:48.451 "is_configured": true, 00:15:48.451 "data_offset": 256, 00:15:48.451 "data_size": 7936 00:15:48.451 }, 00:15:48.451 { 00:15:48.451 "name": "pt2", 00:15:48.451 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:48.451 "is_configured": true, 00:15:48.451 "data_offset": 256, 00:15:48.451 "data_size": 7936 00:15:48.451 } 00:15:48.451 ] 00:15:48.451 } 00:15:48.451 } 00:15:48.451 }' 00:15:48.451 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:48.451 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:15:48.451 pt2' 00:15:48.451 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:48.451 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:15:48.451 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:48.451 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:15:48.451 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:48.451 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.451 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:48.451 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.451 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:15:48.451 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:15:48.451 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:48.451 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:15:48.451 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:48.451 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.451 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:48.451 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:48.711 [2024-11-27 15:20:16.583615] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=af8052d6-b181-460d-9bf2-b55908f07945 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@436 -- # '[' -z af8052d6-b181-460d-9bf2-b55908f07945 ']' 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:48.711 [2024-11-27 15:20:16.635297] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:48.711 [2024-11-27 15:20:16.635376] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:48.711 [2024-11-27 15:20:16.635469] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:48.711 [2024-11-27 15:20:16.635528] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:48.711 [2024-11-27 15:20:16.635537] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@652 -- # local es=0 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:48.711 [2024-11-27 15:20:16.771087] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:15:48.711 [2024-11-27 15:20:16.773178] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:15:48.711 [2024-11-27 15:20:16.773284] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:15:48.711 [2024-11-27 15:20:16.773363] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:15:48.711 [2024-11-27 15:20:16.773408] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:48.711 [2024-11-27 15:20:16.773442] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state configuring 00:15:48.711 request: 00:15:48.711 { 00:15:48.711 "name": "raid_bdev1", 00:15:48.711 "raid_level": "raid1", 00:15:48.711 "base_bdevs": [ 00:15:48.711 "malloc1", 00:15:48.711 "malloc2" 00:15:48.711 ], 00:15:48.711 "superblock": false, 00:15:48.711 "method": "bdev_raid_create", 00:15:48.711 "req_id": 1 00:15:48.711 } 00:15:48.711 Got JSON-RPC error response 00:15:48.711 response: 00:15:48.711 { 00:15:48.711 "code": -17, 00:15:48.711 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:15:48.711 } 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@655 -- # es=1 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:15:48.711 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.970 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:15:48.970 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:15:48.970 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:48.970 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.970 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:48.970 [2024-11-27 15:20:16.835026] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:48.970 [2024-11-27 15:20:16.835124] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:48.970 [2024-11-27 15:20:16.835162] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:15:48.970 [2024-11-27 15:20:16.835205] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:48.970 [2024-11-27 15:20:16.837561] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:48.970 [2024-11-27 15:20:16.837638] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:48.970 [2024-11-27 15:20:16.837723] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:15:48.970 [2024-11-27 15:20:16.837772] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:48.970 pt1 00:15:48.970 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.970 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:15:48.970 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:48.970 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:48.970 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:48.970 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:48.970 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:48.970 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:48.970 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:48.970 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:48.970 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:48.970 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:48.970 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:48.970 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.970 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:48.970 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.970 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:48.970 "name": "raid_bdev1", 00:15:48.970 "uuid": "af8052d6-b181-460d-9bf2-b55908f07945", 00:15:48.970 "strip_size_kb": 0, 00:15:48.970 "state": "configuring", 00:15:48.970 "raid_level": "raid1", 00:15:48.970 "superblock": true, 00:15:48.970 "num_base_bdevs": 2, 00:15:48.970 "num_base_bdevs_discovered": 1, 00:15:48.970 "num_base_bdevs_operational": 2, 00:15:48.970 "base_bdevs_list": [ 00:15:48.970 { 00:15:48.970 "name": "pt1", 00:15:48.970 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:48.970 "is_configured": true, 00:15:48.970 "data_offset": 256, 00:15:48.970 "data_size": 7936 00:15:48.970 }, 00:15:48.970 { 00:15:48.970 "name": null, 00:15:48.970 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:48.970 "is_configured": false, 00:15:48.970 "data_offset": 256, 00:15:48.970 "data_size": 7936 00:15:48.970 } 00:15:48.970 ] 00:15:48.970 }' 00:15:48.970 15:20:16 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:48.970 15:20:16 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:49.230 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:15:49.231 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:15:49.231 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:49.231 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:49.231 15:20:17 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:49.231 15:20:17 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:49.231 [2024-11-27 15:20:17.318219] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:49.231 [2024-11-27 15:20:17.318323] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:49.231 [2024-11-27 15:20:17.318361] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:15:49.231 [2024-11-27 15:20:17.318392] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:49.231 [2024-11-27 15:20:17.318761] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:49.231 [2024-11-27 15:20:17.318824] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:49.231 [2024-11-27 15:20:17.318918] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:49.231 [2024-11-27 15:20:17.318974] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:49.231 [2024-11-27 15:20:17.319076] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:15:49.231 [2024-11-27 15:20:17.319116] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:49.231 [2024-11-27 15:20:17.319369] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:15:49.231 [2024-11-27 15:20:17.319549] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:15:49.231 [2024-11-27 15:20:17.319600] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:15:49.231 [2024-11-27 15:20:17.319740] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:49.231 pt2 00:15:49.231 15:20:17 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:49.231 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:15:49.231 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:49.231 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:49.231 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:49.231 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:49.231 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:49.231 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:49.231 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:49.231 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:49.231 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:49.231 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:49.231 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:49.231 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:49.231 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:49.231 15:20:17 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:49.231 15:20:17 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:49.490 15:20:17 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:49.490 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:49.490 "name": "raid_bdev1", 00:15:49.490 "uuid": "af8052d6-b181-460d-9bf2-b55908f07945", 00:15:49.490 "strip_size_kb": 0, 00:15:49.490 "state": "online", 00:15:49.490 "raid_level": "raid1", 00:15:49.490 "superblock": true, 00:15:49.490 "num_base_bdevs": 2, 00:15:49.490 "num_base_bdevs_discovered": 2, 00:15:49.490 "num_base_bdevs_operational": 2, 00:15:49.490 "base_bdevs_list": [ 00:15:49.490 { 00:15:49.490 "name": "pt1", 00:15:49.490 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:49.490 "is_configured": true, 00:15:49.490 "data_offset": 256, 00:15:49.490 "data_size": 7936 00:15:49.490 }, 00:15:49.490 { 00:15:49.490 "name": "pt2", 00:15:49.490 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:49.490 "is_configured": true, 00:15:49.490 "data_offset": 256, 00:15:49.490 "data_size": 7936 00:15:49.490 } 00:15:49.490 ] 00:15:49.490 }' 00:15:49.490 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:49.490 15:20:17 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:49.750 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:15:49.750 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:15:49.750 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:49.750 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:49.750 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@184 -- # local name 00:15:49.750 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:49.750 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:49.750 15:20:17 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:49.750 15:20:17 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:49.750 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:49.750 [2024-11-27 15:20:17.769722] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:49.750 15:20:17 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:49.750 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:49.750 "name": "raid_bdev1", 00:15:49.750 "aliases": [ 00:15:49.750 "af8052d6-b181-460d-9bf2-b55908f07945" 00:15:49.750 ], 00:15:49.750 "product_name": "Raid Volume", 00:15:49.750 "block_size": 4096, 00:15:49.750 "num_blocks": 7936, 00:15:49.750 "uuid": "af8052d6-b181-460d-9bf2-b55908f07945", 00:15:49.750 "assigned_rate_limits": { 00:15:49.750 "rw_ios_per_sec": 0, 00:15:49.750 "rw_mbytes_per_sec": 0, 00:15:49.750 "r_mbytes_per_sec": 0, 00:15:49.750 "w_mbytes_per_sec": 0 00:15:49.750 }, 00:15:49.750 "claimed": false, 00:15:49.750 "zoned": false, 00:15:49.750 "supported_io_types": { 00:15:49.750 "read": true, 00:15:49.750 "write": true, 00:15:49.750 "unmap": false, 00:15:49.750 "flush": false, 00:15:49.750 "reset": true, 00:15:49.750 "nvme_admin": false, 00:15:49.750 "nvme_io": false, 00:15:49.750 "nvme_io_md": false, 00:15:49.750 "write_zeroes": true, 00:15:49.750 "zcopy": false, 00:15:49.750 "get_zone_info": false, 00:15:49.750 "zone_management": false, 00:15:49.750 "zone_append": false, 00:15:49.750 "compare": false, 00:15:49.750 "compare_and_write": false, 00:15:49.750 "abort": false, 00:15:49.750 "seek_hole": false, 00:15:49.750 "seek_data": false, 00:15:49.750 "copy": false, 00:15:49.750 "nvme_iov_md": false 00:15:49.750 }, 00:15:49.750 "memory_domains": [ 00:15:49.750 { 00:15:49.750 "dma_device_id": "system", 00:15:49.750 "dma_device_type": 1 00:15:49.750 }, 00:15:49.750 { 00:15:49.750 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:49.750 "dma_device_type": 2 00:15:49.750 }, 00:15:49.750 { 00:15:49.750 "dma_device_id": "system", 00:15:49.750 "dma_device_type": 1 00:15:49.750 }, 00:15:49.750 { 00:15:49.750 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:49.750 "dma_device_type": 2 00:15:49.750 } 00:15:49.750 ], 00:15:49.750 "driver_specific": { 00:15:49.750 "raid": { 00:15:49.750 "uuid": "af8052d6-b181-460d-9bf2-b55908f07945", 00:15:49.750 "strip_size_kb": 0, 00:15:49.750 "state": "online", 00:15:49.750 "raid_level": "raid1", 00:15:49.750 "superblock": true, 00:15:49.750 "num_base_bdevs": 2, 00:15:49.750 "num_base_bdevs_discovered": 2, 00:15:49.750 "num_base_bdevs_operational": 2, 00:15:49.750 "base_bdevs_list": [ 00:15:49.750 { 00:15:49.750 "name": "pt1", 00:15:49.750 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:49.750 "is_configured": true, 00:15:49.750 "data_offset": 256, 00:15:49.750 "data_size": 7936 00:15:49.750 }, 00:15:49.750 { 00:15:49.750 "name": "pt2", 00:15:49.750 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:49.750 "is_configured": true, 00:15:49.750 "data_offset": 256, 00:15:49.750 "data_size": 7936 00:15:49.750 } 00:15:49.750 ] 00:15:49.750 } 00:15:49.750 } 00:15:49.750 }' 00:15:49.750 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:49.750 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:15:49.750 pt2' 00:15:49.750 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:50.010 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:15:50.010 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:50.010 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:15:50.010 15:20:17 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.010 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:50.010 15:20:17 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:50.010 15:20:17 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.010 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:15:50.010 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:15:50.010 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:50.010 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:15:50.010 15:20:17 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.010 15:20:17 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:50.010 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:50.010 15:20:17 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.010 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:15:50.010 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:15:50.010 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:50.010 15:20:17 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:15:50.010 15:20:17 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.010 15:20:17 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:50.010 [2024-11-27 15:20:18.005315] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:50.010 15:20:18 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.010 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # '[' af8052d6-b181-460d-9bf2-b55908f07945 '!=' af8052d6-b181-460d-9bf2-b55908f07945 ']' 00:15:50.010 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:15:50.010 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@198 -- # case $1 in 00:15:50.010 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@199 -- # return 0 00:15:50.010 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:15:50.010 15:20:18 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.010 15:20:18 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:50.010 [2024-11-27 15:20:18.033071] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:15:50.010 15:20:18 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.010 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:50.010 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:50.010 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:50.010 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:50.010 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:50.010 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:50.010 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:50.010 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:50.010 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:50.010 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:50.010 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:50.010 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:50.010 15:20:18 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.010 15:20:18 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:50.010 15:20:18 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.010 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:50.010 "name": "raid_bdev1", 00:15:50.010 "uuid": "af8052d6-b181-460d-9bf2-b55908f07945", 00:15:50.010 "strip_size_kb": 0, 00:15:50.010 "state": "online", 00:15:50.010 "raid_level": "raid1", 00:15:50.010 "superblock": true, 00:15:50.010 "num_base_bdevs": 2, 00:15:50.010 "num_base_bdevs_discovered": 1, 00:15:50.010 "num_base_bdevs_operational": 1, 00:15:50.010 "base_bdevs_list": [ 00:15:50.010 { 00:15:50.010 "name": null, 00:15:50.010 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:50.010 "is_configured": false, 00:15:50.010 "data_offset": 0, 00:15:50.010 "data_size": 7936 00:15:50.010 }, 00:15:50.010 { 00:15:50.010 "name": "pt2", 00:15:50.010 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:50.010 "is_configured": true, 00:15:50.010 "data_offset": 256, 00:15:50.010 "data_size": 7936 00:15:50.010 } 00:15:50.010 ] 00:15:50.010 }' 00:15:50.010 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:50.010 15:20:18 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:50.597 [2024-11-27 15:20:18.516218] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:50.597 [2024-11-27 15:20:18.516296] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:50.597 [2024-11-27 15:20:18.516370] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:50.597 [2024-11-27 15:20:18.516427] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:50.597 [2024-11-27 15:20:18.516491] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@519 -- # i=1 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:50.597 [2024-11-27 15:20:18.588096] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:50.597 [2024-11-27 15:20:18.588186] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:50.597 [2024-11-27 15:20:18.588222] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:15:50.597 [2024-11-27 15:20:18.588249] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:50.597 [2024-11-27 15:20:18.590656] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:50.597 [2024-11-27 15:20:18.590732] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:50.597 [2024-11-27 15:20:18.590822] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:50.597 [2024-11-27 15:20:18.590869] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:50.597 [2024-11-27 15:20:18.590983] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:15:50.597 [2024-11-27 15:20:18.591014] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:50.597 [2024-11-27 15:20:18.591260] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:15:50.597 [2024-11-27 15:20:18.591456] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:15:50.597 [2024-11-27 15:20:18.591512] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006d00 00:15:50.597 [2024-11-27 15:20:18.591656] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:50.597 pt2 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:50.597 "name": "raid_bdev1", 00:15:50.597 "uuid": "af8052d6-b181-460d-9bf2-b55908f07945", 00:15:50.597 "strip_size_kb": 0, 00:15:50.597 "state": "online", 00:15:50.597 "raid_level": "raid1", 00:15:50.597 "superblock": true, 00:15:50.597 "num_base_bdevs": 2, 00:15:50.597 "num_base_bdevs_discovered": 1, 00:15:50.597 "num_base_bdevs_operational": 1, 00:15:50.597 "base_bdevs_list": [ 00:15:50.597 { 00:15:50.597 "name": null, 00:15:50.597 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:50.597 "is_configured": false, 00:15:50.597 "data_offset": 256, 00:15:50.597 "data_size": 7936 00:15:50.597 }, 00:15:50.597 { 00:15:50.597 "name": "pt2", 00:15:50.597 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:50.597 "is_configured": true, 00:15:50.597 "data_offset": 256, 00:15:50.597 "data_size": 7936 00:15:50.597 } 00:15:50.597 ] 00:15:50.597 }' 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:50.597 15:20:18 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:51.182 15:20:19 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:51.182 15:20:19 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.182 15:20:19 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:51.182 [2024-11-27 15:20:19.015529] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:51.182 [2024-11-27 15:20:19.015606] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:51.182 [2024-11-27 15:20:19.015682] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:51.182 [2024-11-27 15:20:19.015736] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:51.182 [2024-11-27 15:20:19.015805] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name raid_bdev1, state offline 00:15:51.182 15:20:19 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.182 15:20:19 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:15:51.182 15:20:19 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:51.182 15:20:19 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.182 15:20:19 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:51.182 15:20:19 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.182 15:20:19 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:15:51.182 15:20:19 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:15:51.183 15:20:19 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:15:51.183 15:20:19 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:51.183 15:20:19 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.183 15:20:19 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:51.183 [2024-11-27 15:20:19.075502] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:51.183 [2024-11-27 15:20:19.075602] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:51.183 [2024-11-27 15:20:19.075635] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:15:51.183 [2024-11-27 15:20:19.075677] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:51.183 [2024-11-27 15:20:19.078019] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:51.183 [2024-11-27 15:20:19.078098] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:51.183 [2024-11-27 15:20:19.078177] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:15:51.183 [2024-11-27 15:20:19.078236] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:51.183 [2024-11-27 15:20:19.078356] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:15:51.183 [2024-11-27 15:20:19.078442] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:51.183 [2024-11-27 15:20:19.078503] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007080 name raid_bdev1, state configuring 00:15:51.183 [2024-11-27 15:20:19.078588] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:51.183 [2024-11-27 15:20:19.078681] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007400 00:15:51.183 [2024-11-27 15:20:19.078725] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:51.183 [2024-11-27 15:20:19.078960] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:15:51.183 [2024-11-27 15:20:19.079120] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007400 00:15:51.183 [2024-11-27 15:20:19.079165] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007400 00:15:51.183 pt1 00:15:51.183 [2024-11-27 15:20:19.079317] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:51.183 15:20:19 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.183 15:20:19 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:15:51.183 15:20:19 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:51.183 15:20:19 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:51.183 15:20:19 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:51.183 15:20:19 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:51.183 15:20:19 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:51.183 15:20:19 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:51.183 15:20:19 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:51.183 15:20:19 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:51.183 15:20:19 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:51.183 15:20:19 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:51.183 15:20:19 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:51.183 15:20:19 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.183 15:20:19 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:51.183 15:20:19 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:51.183 15:20:19 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.183 15:20:19 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:51.183 "name": "raid_bdev1", 00:15:51.183 "uuid": "af8052d6-b181-460d-9bf2-b55908f07945", 00:15:51.183 "strip_size_kb": 0, 00:15:51.183 "state": "online", 00:15:51.183 "raid_level": "raid1", 00:15:51.183 "superblock": true, 00:15:51.183 "num_base_bdevs": 2, 00:15:51.183 "num_base_bdevs_discovered": 1, 00:15:51.183 "num_base_bdevs_operational": 1, 00:15:51.183 "base_bdevs_list": [ 00:15:51.183 { 00:15:51.183 "name": null, 00:15:51.183 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:51.183 "is_configured": false, 00:15:51.183 "data_offset": 256, 00:15:51.183 "data_size": 7936 00:15:51.183 }, 00:15:51.183 { 00:15:51.183 "name": "pt2", 00:15:51.183 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:51.183 "is_configured": true, 00:15:51.183 "data_offset": 256, 00:15:51.183 "data_size": 7936 00:15:51.183 } 00:15:51.183 ] 00:15:51.183 }' 00:15:51.183 15:20:19 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:51.183 15:20:19 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:51.752 15:20:19 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:15:51.752 15:20:19 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:15:51.752 15:20:19 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.752 15:20:19 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:51.752 15:20:19 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.752 15:20:19 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:15:51.752 15:20:19 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:51.752 15:20:19 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:15:51.752 15:20:19 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.752 15:20:19 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:51.752 [2024-11-27 15:20:19.603102] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:51.752 15:20:19 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.752 15:20:19 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # '[' af8052d6-b181-460d-9bf2-b55908f07945 '!=' af8052d6-b181-460d-9bf2-b55908f07945 ']' 00:15:51.752 15:20:19 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@563 -- # killprocess 96754 00:15:51.752 15:20:19 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@954 -- # '[' -z 96754 ']' 00:15:51.752 15:20:19 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@958 -- # kill -0 96754 00:15:51.752 15:20:19 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@959 -- # uname 00:15:51.752 15:20:19 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:51.752 15:20:19 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 96754 00:15:51.752 15:20:19 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:51.752 15:20:19 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:51.752 killing process with pid 96754 00:15:51.752 15:20:19 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@972 -- # echo 'killing process with pid 96754' 00:15:51.752 15:20:19 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@973 -- # kill 96754 00:15:51.752 [2024-11-27 15:20:19.682179] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:51.752 [2024-11-27 15:20:19.682232] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:51.752 [2024-11-27 15:20:19.682267] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:51.752 [2024-11-27 15:20:19.682276] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007400 name raid_bdev1, state offline 00:15:51.752 15:20:19 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@978 -- # wait 96754 00:15:51.752 [2024-11-27 15:20:19.724854] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:52.013 15:20:20 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@565 -- # return 0 00:15:52.013 00:15:52.013 real 0m5.160s 00:15:52.013 user 0m8.329s 00:15:52.013 sys 0m1.100s 00:15:52.013 15:20:20 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:52.013 ************************************ 00:15:52.013 END TEST raid_superblock_test_4k 00:15:52.013 ************************************ 00:15:52.013 15:20:20 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:52.274 15:20:20 bdev_raid -- bdev/bdev_raid.sh@999 -- # '[' true = true ']' 00:15:52.274 15:20:20 bdev_raid -- bdev/bdev_raid.sh@1000 -- # run_test raid_rebuild_test_sb_4k raid_rebuild_test raid1 2 true false true 00:15:52.274 15:20:20 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:15:52.274 15:20:20 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:52.274 15:20:20 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:52.274 ************************************ 00:15:52.274 START TEST raid_rebuild_test_sb_4k 00:15:52.274 ************************************ 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true false true 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@573 -- # local verify=true 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # local strip_size 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@577 -- # local create_arg 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@579 -- # local data_offset 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@597 -- # raid_pid=97071 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@598 -- # waitforlisten 97071 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@835 -- # '[' -z 97071 ']' 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:52.274 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:52.274 15:20:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:52.274 [2024-11-27 15:20:20.245823] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:15:52.274 [2024-11-27 15:20:20.246050] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.ealI/O size of 3145728 is greater than zero copy threshold (65536). 00:15:52.274 Zero copy mechanism will not be used. 00:15:52.274 :6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid97071 ] 00:15:52.535 [2024-11-27 15:20:20.416535] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:52.535 [2024-11-27 15:20:20.456984] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:52.535 [2024-11-27 15:20:20.533427] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:52.535 [2024-11-27 15:20:20.533574] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:53.106 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:53.106 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@868 -- # return 0 00:15:53.106 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev1_malloc 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:53.107 BaseBdev1_malloc 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:53.107 [2024-11-27 15:20:21.090600] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:53.107 [2024-11-27 15:20:21.090734] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:53.107 [2024-11-27 15:20:21.090771] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:53.107 [2024-11-27 15:20:21.090785] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:53.107 [2024-11-27 15:20:21.093326] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:53.107 [2024-11-27 15:20:21.093372] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:53.107 BaseBdev1 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev2_malloc 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:53.107 BaseBdev2_malloc 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:53.107 [2024-11-27 15:20:21.125145] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:15:53.107 [2024-11-27 15:20:21.125267] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:53.107 [2024-11-27 15:20:21.125311] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:15:53.107 [2024-11-27 15:20:21.125343] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:53.107 [2024-11-27 15:20:21.127740] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:53.107 [2024-11-27 15:20:21.127827] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:15:53.107 BaseBdev2 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -b spare_malloc 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:53.107 spare_malloc 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:53.107 spare_delay 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:53.107 [2024-11-27 15:20:21.171638] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:53.107 [2024-11-27 15:20:21.171754] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:53.107 [2024-11-27 15:20:21.171784] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:15:53.107 [2024-11-27 15:20:21.171794] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:53.107 [2024-11-27 15:20:21.174200] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:53.107 [2024-11-27 15:20:21.174241] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:53.107 spare 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:53.107 [2024-11-27 15:20:21.183652] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:53.107 [2024-11-27 15:20:21.185739] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:53.107 [2024-11-27 15:20:21.185979] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:15:53.107 [2024-11-27 15:20:21.186020] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:53.107 [2024-11-27 15:20:21.186329] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:15:53.107 [2024-11-27 15:20:21.186540] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:15:53.107 [2024-11-27 15:20:21.186595] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:15:53.107 [2024-11-27 15:20:21.186769] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.107 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:53.367 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.367 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:53.367 "name": "raid_bdev1", 00:15:53.367 "uuid": "3e22d42e-5cce-482d-b7c7-e6b96a0952e1", 00:15:53.367 "strip_size_kb": 0, 00:15:53.367 "state": "online", 00:15:53.367 "raid_level": "raid1", 00:15:53.367 "superblock": true, 00:15:53.367 "num_base_bdevs": 2, 00:15:53.367 "num_base_bdevs_discovered": 2, 00:15:53.367 "num_base_bdevs_operational": 2, 00:15:53.367 "base_bdevs_list": [ 00:15:53.367 { 00:15:53.367 "name": "BaseBdev1", 00:15:53.367 "uuid": "4b77d2cc-c324-5fef-94a8-93d7f4ee020a", 00:15:53.367 "is_configured": true, 00:15:53.367 "data_offset": 256, 00:15:53.367 "data_size": 7936 00:15:53.367 }, 00:15:53.367 { 00:15:53.367 "name": "BaseBdev2", 00:15:53.367 "uuid": "13d2d49d-c27f-5472-8a1a-07c03d9e4fee", 00:15:53.367 "is_configured": true, 00:15:53.367 "data_offset": 256, 00:15:53.367 "data_size": 7936 00:15:53.367 } 00:15:53.367 ] 00:15:53.367 }' 00:15:53.367 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:53.367 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:53.627 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:15:53.627 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:53.627 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.627 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:53.627 [2024-11-27 15:20:21.599710] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:53.627 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.627 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:15:53.627 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:53.627 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.627 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:15:53.627 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:53.627 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.627 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:15:53.627 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:15:53.627 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:15:53.628 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:15:53.628 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:15:53.628 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:53.628 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:15:53.628 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:53.628 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:15:53.628 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:53.628 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@12 -- # local i 00:15:53.628 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:53.628 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:53.628 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:15:53.887 [2024-11-27 15:20:21.871055] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:15:53.887 /dev/nbd0 00:15:53.887 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:53.887 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:53.887 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:15:53.887 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # local i 00:15:53.887 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:53.887 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:53.887 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:15:53.887 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@877 -- # break 00:15:53.887 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:53.887 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:53.887 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:53.887 1+0 records in 00:15:53.887 1+0 records out 00:15:53.887 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000464327 s, 8.8 MB/s 00:15:53.887 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:53.887 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # size=4096 00:15:53.887 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:53.887 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:53.887 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@893 -- # return 0 00:15:53.887 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:53.887 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:53.887 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:15:53.887 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:15:53.887 15:20:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=4096 count=7936 oflag=direct 00:15:54.824 7936+0 records in 00:15:54.824 7936+0 records out 00:15:54.824 32505856 bytes (33 MB, 31 MiB) copied, 0.622738 s, 52.2 MB/s 00:15:54.824 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:15:54.824 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:54.824 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:15:54.824 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:54.824 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@51 -- # local i 00:15:54.824 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:54.824 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:54.824 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:54.824 [2024-11-27 15:20:22.788952] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:54.824 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:54.824 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:54.824 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:54.824 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:54.824 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:54.824 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:15:54.824 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:15:54.824 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:15:54.824 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:54.825 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:54.825 [2024-11-27 15:20:22.801518] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:54.825 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:54.825 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:54.825 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:54.825 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:54.825 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:54.825 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:54.825 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:54.825 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:54.825 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:54.825 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:54.825 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:54.825 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:54.825 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:54.825 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:54.825 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:54.825 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:54.825 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:54.825 "name": "raid_bdev1", 00:15:54.825 "uuid": "3e22d42e-5cce-482d-b7c7-e6b96a0952e1", 00:15:54.825 "strip_size_kb": 0, 00:15:54.825 "state": "online", 00:15:54.825 "raid_level": "raid1", 00:15:54.825 "superblock": true, 00:15:54.825 "num_base_bdevs": 2, 00:15:54.825 "num_base_bdevs_discovered": 1, 00:15:54.825 "num_base_bdevs_operational": 1, 00:15:54.825 "base_bdevs_list": [ 00:15:54.825 { 00:15:54.825 "name": null, 00:15:54.825 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:54.825 "is_configured": false, 00:15:54.825 "data_offset": 0, 00:15:54.825 "data_size": 7936 00:15:54.825 }, 00:15:54.825 { 00:15:54.825 "name": "BaseBdev2", 00:15:54.825 "uuid": "13d2d49d-c27f-5472-8a1a-07c03d9e4fee", 00:15:54.825 "is_configured": true, 00:15:54.825 "data_offset": 256, 00:15:54.825 "data_size": 7936 00:15:54.825 } 00:15:54.825 ] 00:15:54.825 }' 00:15:54.825 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:54.825 15:20:22 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:55.394 15:20:23 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:55.394 15:20:23 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:55.394 15:20:23 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:55.394 [2024-11-27 15:20:23.212845] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:55.394 [2024-11-27 15:20:23.218116] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00018d0c0 00:15:55.394 15:20:23 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:55.394 15:20:23 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@647 -- # sleep 1 00:15:55.394 [2024-11-27 15:20:23.220168] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:56.333 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:56.333 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:56.333 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:56.333 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:56.333 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:56.333 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:56.333 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:56.333 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:56.333 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:56.333 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:56.334 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:56.334 "name": "raid_bdev1", 00:15:56.334 "uuid": "3e22d42e-5cce-482d-b7c7-e6b96a0952e1", 00:15:56.334 "strip_size_kb": 0, 00:15:56.334 "state": "online", 00:15:56.334 "raid_level": "raid1", 00:15:56.334 "superblock": true, 00:15:56.334 "num_base_bdevs": 2, 00:15:56.334 "num_base_bdevs_discovered": 2, 00:15:56.334 "num_base_bdevs_operational": 2, 00:15:56.334 "process": { 00:15:56.334 "type": "rebuild", 00:15:56.334 "target": "spare", 00:15:56.334 "progress": { 00:15:56.334 "blocks": 2560, 00:15:56.334 "percent": 32 00:15:56.334 } 00:15:56.334 }, 00:15:56.334 "base_bdevs_list": [ 00:15:56.334 { 00:15:56.334 "name": "spare", 00:15:56.334 "uuid": "2a0b07f9-d30d-52b2-bf69-c956b4c05376", 00:15:56.334 "is_configured": true, 00:15:56.334 "data_offset": 256, 00:15:56.334 "data_size": 7936 00:15:56.334 }, 00:15:56.334 { 00:15:56.334 "name": "BaseBdev2", 00:15:56.334 "uuid": "13d2d49d-c27f-5472-8a1a-07c03d9e4fee", 00:15:56.334 "is_configured": true, 00:15:56.334 "data_offset": 256, 00:15:56.334 "data_size": 7936 00:15:56.334 } 00:15:56.334 ] 00:15:56.334 }' 00:15:56.334 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:56.334 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:56.334 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:56.334 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:56.334 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:56.334 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:56.334 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:56.334 [2024-11-27 15:20:24.379825] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:56.334 [2024-11-27 15:20:24.424741] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:56.334 [2024-11-27 15:20:24.424863] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:56.334 [2024-11-27 15:20:24.424911] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:56.334 [2024-11-27 15:20:24.424950] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:56.334 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:56.334 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:56.334 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:56.334 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:56.334 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:56.334 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:56.334 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:56.594 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:56.594 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:56.594 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:56.594 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:56.594 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:56.594 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:56.594 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:56.594 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:56.594 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:56.594 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:56.594 "name": "raid_bdev1", 00:15:56.595 "uuid": "3e22d42e-5cce-482d-b7c7-e6b96a0952e1", 00:15:56.595 "strip_size_kb": 0, 00:15:56.595 "state": "online", 00:15:56.595 "raid_level": "raid1", 00:15:56.595 "superblock": true, 00:15:56.595 "num_base_bdevs": 2, 00:15:56.595 "num_base_bdevs_discovered": 1, 00:15:56.595 "num_base_bdevs_operational": 1, 00:15:56.595 "base_bdevs_list": [ 00:15:56.595 { 00:15:56.595 "name": null, 00:15:56.595 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:56.595 "is_configured": false, 00:15:56.595 "data_offset": 0, 00:15:56.595 "data_size": 7936 00:15:56.595 }, 00:15:56.595 { 00:15:56.595 "name": "BaseBdev2", 00:15:56.595 "uuid": "13d2d49d-c27f-5472-8a1a-07c03d9e4fee", 00:15:56.595 "is_configured": true, 00:15:56.595 "data_offset": 256, 00:15:56.595 "data_size": 7936 00:15:56.595 } 00:15:56.595 ] 00:15:56.595 }' 00:15:56.595 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:56.595 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:56.855 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:56.855 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:56.855 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:56.855 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:56.855 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:56.855 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:56.855 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:56.855 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:56.855 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:56.855 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:56.855 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:56.855 "name": "raid_bdev1", 00:15:56.855 "uuid": "3e22d42e-5cce-482d-b7c7-e6b96a0952e1", 00:15:56.855 "strip_size_kb": 0, 00:15:56.855 "state": "online", 00:15:56.855 "raid_level": "raid1", 00:15:56.855 "superblock": true, 00:15:56.855 "num_base_bdevs": 2, 00:15:56.855 "num_base_bdevs_discovered": 1, 00:15:56.855 "num_base_bdevs_operational": 1, 00:15:56.855 "base_bdevs_list": [ 00:15:56.855 { 00:15:56.855 "name": null, 00:15:56.855 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:56.855 "is_configured": false, 00:15:56.855 "data_offset": 0, 00:15:56.855 "data_size": 7936 00:15:56.855 }, 00:15:56.855 { 00:15:56.855 "name": "BaseBdev2", 00:15:56.855 "uuid": "13d2d49d-c27f-5472-8a1a-07c03d9e4fee", 00:15:56.855 "is_configured": true, 00:15:56.855 "data_offset": 256, 00:15:56.855 "data_size": 7936 00:15:56.855 } 00:15:56.855 ] 00:15:56.855 }' 00:15:56.855 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:57.115 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:57.115 15:20:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:57.115 15:20:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:57.115 15:20:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:57.115 15:20:25 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:57.115 15:20:25 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:57.115 [2024-11-27 15:20:25.013033] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:57.115 [2024-11-27 15:20:25.017906] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00018d190 00:15:57.115 15:20:25 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:57.115 15:20:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@663 -- # sleep 1 00:15:57.115 [2024-11-27 15:20:25.019836] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:58.054 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:58.054 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:58.054 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:58.054 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:58.054 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:58.054 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:58.054 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:58.054 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.054 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:58.054 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.054 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:58.054 "name": "raid_bdev1", 00:15:58.054 "uuid": "3e22d42e-5cce-482d-b7c7-e6b96a0952e1", 00:15:58.054 "strip_size_kb": 0, 00:15:58.054 "state": "online", 00:15:58.054 "raid_level": "raid1", 00:15:58.054 "superblock": true, 00:15:58.054 "num_base_bdevs": 2, 00:15:58.054 "num_base_bdevs_discovered": 2, 00:15:58.054 "num_base_bdevs_operational": 2, 00:15:58.054 "process": { 00:15:58.054 "type": "rebuild", 00:15:58.054 "target": "spare", 00:15:58.054 "progress": { 00:15:58.054 "blocks": 2560, 00:15:58.054 "percent": 32 00:15:58.054 } 00:15:58.054 }, 00:15:58.054 "base_bdevs_list": [ 00:15:58.054 { 00:15:58.054 "name": "spare", 00:15:58.054 "uuid": "2a0b07f9-d30d-52b2-bf69-c956b4c05376", 00:15:58.054 "is_configured": true, 00:15:58.054 "data_offset": 256, 00:15:58.055 "data_size": 7936 00:15:58.055 }, 00:15:58.055 { 00:15:58.055 "name": "BaseBdev2", 00:15:58.055 "uuid": "13d2d49d-c27f-5472-8a1a-07c03d9e4fee", 00:15:58.055 "is_configured": true, 00:15:58.055 "data_offset": 256, 00:15:58.055 "data_size": 7936 00:15:58.055 } 00:15:58.055 ] 00:15:58.055 }' 00:15:58.055 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:58.055 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:58.055 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:58.055 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:58.055 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:15:58.055 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:15:58.055 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:15:58.055 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:15:58.055 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:15:58.055 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:15:58.055 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@706 -- # local timeout=567 00:15:58.055 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:58.055 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:58.055 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:58.055 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:58.055 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:58.055 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:58.315 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:58.315 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:58.315 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.315 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:58.315 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.315 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:58.315 "name": "raid_bdev1", 00:15:58.315 "uuid": "3e22d42e-5cce-482d-b7c7-e6b96a0952e1", 00:15:58.315 "strip_size_kb": 0, 00:15:58.315 "state": "online", 00:15:58.315 "raid_level": "raid1", 00:15:58.315 "superblock": true, 00:15:58.315 "num_base_bdevs": 2, 00:15:58.315 "num_base_bdevs_discovered": 2, 00:15:58.315 "num_base_bdevs_operational": 2, 00:15:58.315 "process": { 00:15:58.315 "type": "rebuild", 00:15:58.315 "target": "spare", 00:15:58.315 "progress": { 00:15:58.315 "blocks": 2816, 00:15:58.315 "percent": 35 00:15:58.315 } 00:15:58.315 }, 00:15:58.315 "base_bdevs_list": [ 00:15:58.315 { 00:15:58.315 "name": "spare", 00:15:58.315 "uuid": "2a0b07f9-d30d-52b2-bf69-c956b4c05376", 00:15:58.315 "is_configured": true, 00:15:58.315 "data_offset": 256, 00:15:58.315 "data_size": 7936 00:15:58.315 }, 00:15:58.315 { 00:15:58.315 "name": "BaseBdev2", 00:15:58.316 "uuid": "13d2d49d-c27f-5472-8a1a-07c03d9e4fee", 00:15:58.316 "is_configured": true, 00:15:58.316 "data_offset": 256, 00:15:58.316 "data_size": 7936 00:15:58.316 } 00:15:58.316 ] 00:15:58.316 }' 00:15:58.316 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:58.316 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:58.316 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:58.316 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:58.316 15:20:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:59.255 15:20:27 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:59.256 15:20:27 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:59.256 15:20:27 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:59.256 15:20:27 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:59.256 15:20:27 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:59.256 15:20:27 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:59.256 15:20:27 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:59.256 15:20:27 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.256 15:20:27 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:59.256 15:20:27 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:59.256 15:20:27 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.256 15:20:27 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:59.256 "name": "raid_bdev1", 00:15:59.256 "uuid": "3e22d42e-5cce-482d-b7c7-e6b96a0952e1", 00:15:59.256 "strip_size_kb": 0, 00:15:59.256 "state": "online", 00:15:59.256 "raid_level": "raid1", 00:15:59.256 "superblock": true, 00:15:59.256 "num_base_bdevs": 2, 00:15:59.256 "num_base_bdevs_discovered": 2, 00:15:59.256 "num_base_bdevs_operational": 2, 00:15:59.256 "process": { 00:15:59.256 "type": "rebuild", 00:15:59.256 "target": "spare", 00:15:59.256 "progress": { 00:15:59.256 "blocks": 5632, 00:15:59.256 "percent": 70 00:15:59.256 } 00:15:59.256 }, 00:15:59.256 "base_bdevs_list": [ 00:15:59.256 { 00:15:59.256 "name": "spare", 00:15:59.256 "uuid": "2a0b07f9-d30d-52b2-bf69-c956b4c05376", 00:15:59.256 "is_configured": true, 00:15:59.256 "data_offset": 256, 00:15:59.256 "data_size": 7936 00:15:59.256 }, 00:15:59.256 { 00:15:59.256 "name": "BaseBdev2", 00:15:59.256 "uuid": "13d2d49d-c27f-5472-8a1a-07c03d9e4fee", 00:15:59.256 "is_configured": true, 00:15:59.256 "data_offset": 256, 00:15:59.256 "data_size": 7936 00:15:59.256 } 00:15:59.256 ] 00:15:59.256 }' 00:15:59.516 15:20:27 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:59.516 15:20:27 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:59.516 15:20:27 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:59.516 15:20:27 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:59.516 15:20:27 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:00.085 [2024-11-27 15:20:28.130329] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:16:00.085 [2024-11-27 15:20:28.130402] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:16:00.085 [2024-11-27 15:20:28.130499] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:00.344 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:00.344 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:00.344 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:00.344 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:00.344 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:00.344 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:00.344 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:00.344 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:00.344 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.344 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:00.604 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.604 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:00.604 "name": "raid_bdev1", 00:16:00.604 "uuid": "3e22d42e-5cce-482d-b7c7-e6b96a0952e1", 00:16:00.604 "strip_size_kb": 0, 00:16:00.604 "state": "online", 00:16:00.604 "raid_level": "raid1", 00:16:00.604 "superblock": true, 00:16:00.604 "num_base_bdevs": 2, 00:16:00.604 "num_base_bdevs_discovered": 2, 00:16:00.604 "num_base_bdevs_operational": 2, 00:16:00.604 "base_bdevs_list": [ 00:16:00.604 { 00:16:00.604 "name": "spare", 00:16:00.604 "uuid": "2a0b07f9-d30d-52b2-bf69-c956b4c05376", 00:16:00.604 "is_configured": true, 00:16:00.604 "data_offset": 256, 00:16:00.604 "data_size": 7936 00:16:00.604 }, 00:16:00.604 { 00:16:00.604 "name": "BaseBdev2", 00:16:00.604 "uuid": "13d2d49d-c27f-5472-8a1a-07c03d9e4fee", 00:16:00.604 "is_configured": true, 00:16:00.604 "data_offset": 256, 00:16:00.604 "data_size": 7936 00:16:00.604 } 00:16:00.604 ] 00:16:00.604 }' 00:16:00.604 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:00.604 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:16:00.604 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:00.604 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:16:00.604 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@709 -- # break 00:16:00.604 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:00.604 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:00.604 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:00.604 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:00.604 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:00.604 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:00.604 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:00.604 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.604 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:00.604 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.604 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:00.604 "name": "raid_bdev1", 00:16:00.604 "uuid": "3e22d42e-5cce-482d-b7c7-e6b96a0952e1", 00:16:00.604 "strip_size_kb": 0, 00:16:00.604 "state": "online", 00:16:00.604 "raid_level": "raid1", 00:16:00.604 "superblock": true, 00:16:00.604 "num_base_bdevs": 2, 00:16:00.604 "num_base_bdevs_discovered": 2, 00:16:00.604 "num_base_bdevs_operational": 2, 00:16:00.604 "base_bdevs_list": [ 00:16:00.604 { 00:16:00.604 "name": "spare", 00:16:00.604 "uuid": "2a0b07f9-d30d-52b2-bf69-c956b4c05376", 00:16:00.604 "is_configured": true, 00:16:00.604 "data_offset": 256, 00:16:00.604 "data_size": 7936 00:16:00.604 }, 00:16:00.604 { 00:16:00.604 "name": "BaseBdev2", 00:16:00.604 "uuid": "13d2d49d-c27f-5472-8a1a-07c03d9e4fee", 00:16:00.604 "is_configured": true, 00:16:00.604 "data_offset": 256, 00:16:00.604 "data_size": 7936 00:16:00.604 } 00:16:00.604 ] 00:16:00.604 }' 00:16:00.604 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:00.604 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:00.604 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:00.864 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:00.864 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:00.864 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:00.864 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:00.864 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:00.864 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:00.864 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:00.864 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:00.864 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:00.864 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:00.864 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:00.864 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:00.864 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:00.864 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.864 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:00.864 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.864 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:00.864 "name": "raid_bdev1", 00:16:00.864 "uuid": "3e22d42e-5cce-482d-b7c7-e6b96a0952e1", 00:16:00.864 "strip_size_kb": 0, 00:16:00.864 "state": "online", 00:16:00.864 "raid_level": "raid1", 00:16:00.864 "superblock": true, 00:16:00.864 "num_base_bdevs": 2, 00:16:00.864 "num_base_bdevs_discovered": 2, 00:16:00.864 "num_base_bdevs_operational": 2, 00:16:00.864 "base_bdevs_list": [ 00:16:00.864 { 00:16:00.864 "name": "spare", 00:16:00.864 "uuid": "2a0b07f9-d30d-52b2-bf69-c956b4c05376", 00:16:00.864 "is_configured": true, 00:16:00.864 "data_offset": 256, 00:16:00.864 "data_size": 7936 00:16:00.864 }, 00:16:00.864 { 00:16:00.864 "name": "BaseBdev2", 00:16:00.864 "uuid": "13d2d49d-c27f-5472-8a1a-07c03d9e4fee", 00:16:00.864 "is_configured": true, 00:16:00.864 "data_offset": 256, 00:16:00.864 "data_size": 7936 00:16:00.864 } 00:16:00.864 ] 00:16:00.864 }' 00:16:00.864 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:00.864 15:20:28 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:01.125 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:01.125 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:01.125 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:01.125 [2024-11-27 15:20:29.177293] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:01.125 [2024-11-27 15:20:29.177384] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:01.125 [2024-11-27 15:20:29.177503] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:01.125 [2024-11-27 15:20:29.177611] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:01.125 [2024-11-27 15:20:29.177659] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:16:01.125 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:01.125 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:01.125 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # jq length 00:16:01.125 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:01.125 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:01.125 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:01.385 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:16:01.385 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:16:01.385 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:16:01.385 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:16:01.385 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:16:01.385 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:16:01.385 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # local bdev_list 00:16:01.385 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:16:01.385 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # local nbd_list 00:16:01.385 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@12 -- # local i 00:16:01.385 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:16:01.385 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:01.385 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:16:01.385 /dev/nbd0 00:16:01.385 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:16:01.385 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:16:01.385 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:16:01.385 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # local i 00:16:01.385 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:16:01.385 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:16:01.385 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:16:01.385 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@877 -- # break 00:16:01.385 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:16:01.385 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:16:01.385 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:01.385 1+0 records in 00:16:01.385 1+0 records out 00:16:01.385 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000492888 s, 8.3 MB/s 00:16:01.385 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:01.645 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # size=4096 00:16:01.645 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:01.645 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:16:01.645 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@893 -- # return 0 00:16:01.645 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:01.645 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:01.645 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:16:01.645 /dev/nbd1 00:16:01.645 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:16:01.645 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:16:01.645 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:16:01.645 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # local i 00:16:01.645 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:16:01.645 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:16:01.645 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:16:01.645 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@877 -- # break 00:16:01.645 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:16:01.645 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:16:01.645 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:01.645 1+0 records in 00:16:01.645 1+0 records out 00:16:01.645 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000421689 s, 9.7 MB/s 00:16:01.645 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:01.645 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # size=4096 00:16:01.645 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:01.905 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:16:01.905 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@893 -- # return 0 00:16:01.905 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:01.905 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:01.905 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:16:01.905 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:16:01.905 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:16:01.905 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:16:01.905 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:01.905 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@51 -- # local i 00:16:01.905 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:01.905 15:20:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:16:02.166 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:02.166 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:02.166 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:02.166 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:02.166 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:02.166 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:02.166 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:16:02.166 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:16:02.166 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:02.166 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:16:02.166 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:16:02.166 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:16:02.166 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:16:02.166 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:02.166 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:02.166 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:16:02.166 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:16:02.166 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:16:02.166 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:16:02.166 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:16:02.166 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:02.166 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:02.166 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:02.166 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:02.166 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:02.166 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:02.166 [2024-11-27 15:20:30.254174] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:02.166 [2024-11-27 15:20:30.254256] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:02.166 [2024-11-27 15:20:30.254278] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:16:02.166 [2024-11-27 15:20:30.254291] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:02.166 [2024-11-27 15:20:30.256356] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:02.166 [2024-11-27 15:20:30.256397] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:02.166 [2024-11-27 15:20:30.256480] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:16:02.166 [2024-11-27 15:20:30.256520] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:02.166 [2024-11-27 15:20:30.256648] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:02.166 spare 00:16:02.166 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:02.166 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:16:02.166 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:02.166 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:02.427 [2024-11-27 15:20:30.356564] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006600 00:16:02.427 [2024-11-27 15:20:30.356587] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:02.427 [2024-11-27 15:20:30.356849] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c19b0 00:16:02.427 [2024-11-27 15:20:30.357030] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006600 00:16:02.427 [2024-11-27 15:20:30.357044] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006600 00:16:02.427 [2024-11-27 15:20:30.357173] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:02.427 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:02.427 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:02.427 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:02.427 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:02.427 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:02.427 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:02.427 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:02.427 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:02.427 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:02.427 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:02.427 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:02.427 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:02.427 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:02.427 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:02.427 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:02.427 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:02.427 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:02.427 "name": "raid_bdev1", 00:16:02.427 "uuid": "3e22d42e-5cce-482d-b7c7-e6b96a0952e1", 00:16:02.427 "strip_size_kb": 0, 00:16:02.427 "state": "online", 00:16:02.427 "raid_level": "raid1", 00:16:02.427 "superblock": true, 00:16:02.427 "num_base_bdevs": 2, 00:16:02.427 "num_base_bdevs_discovered": 2, 00:16:02.427 "num_base_bdevs_operational": 2, 00:16:02.427 "base_bdevs_list": [ 00:16:02.427 { 00:16:02.427 "name": "spare", 00:16:02.427 "uuid": "2a0b07f9-d30d-52b2-bf69-c956b4c05376", 00:16:02.427 "is_configured": true, 00:16:02.427 "data_offset": 256, 00:16:02.427 "data_size": 7936 00:16:02.427 }, 00:16:02.427 { 00:16:02.427 "name": "BaseBdev2", 00:16:02.427 "uuid": "13d2d49d-c27f-5472-8a1a-07c03d9e4fee", 00:16:02.427 "is_configured": true, 00:16:02.427 "data_offset": 256, 00:16:02.427 "data_size": 7936 00:16:02.427 } 00:16:02.427 ] 00:16:02.427 }' 00:16:02.427 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:02.427 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:02.688 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:02.688 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:02.688 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:02.688 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:02.688 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:02.688 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:02.688 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:02.688 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:02.688 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:02.688 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:02.948 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:02.948 "name": "raid_bdev1", 00:16:02.948 "uuid": "3e22d42e-5cce-482d-b7c7-e6b96a0952e1", 00:16:02.948 "strip_size_kb": 0, 00:16:02.948 "state": "online", 00:16:02.948 "raid_level": "raid1", 00:16:02.948 "superblock": true, 00:16:02.948 "num_base_bdevs": 2, 00:16:02.948 "num_base_bdevs_discovered": 2, 00:16:02.948 "num_base_bdevs_operational": 2, 00:16:02.948 "base_bdevs_list": [ 00:16:02.948 { 00:16:02.948 "name": "spare", 00:16:02.948 "uuid": "2a0b07f9-d30d-52b2-bf69-c956b4c05376", 00:16:02.948 "is_configured": true, 00:16:02.948 "data_offset": 256, 00:16:02.948 "data_size": 7936 00:16:02.948 }, 00:16:02.948 { 00:16:02.948 "name": "BaseBdev2", 00:16:02.948 "uuid": "13d2d49d-c27f-5472-8a1a-07c03d9e4fee", 00:16:02.948 "is_configured": true, 00:16:02.948 "data_offset": 256, 00:16:02.949 "data_size": 7936 00:16:02.949 } 00:16:02.949 ] 00:16:02.949 }' 00:16:02.949 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:02.949 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:02.949 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:02.949 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:02.949 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:02.949 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:16:02.949 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:02.949 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:02.949 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:02.949 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:16:02.949 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:02.949 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:02.949 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:02.949 [2024-11-27 15:20:30.972968] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:02.949 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:02.949 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:02.949 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:02.949 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:02.949 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:02.949 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:02.949 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:02.949 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:02.949 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:02.949 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:02.949 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:02.949 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:02.949 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:02.949 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:02.949 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:02.949 15:20:30 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:02.949 15:20:31 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:02.949 "name": "raid_bdev1", 00:16:02.949 "uuid": "3e22d42e-5cce-482d-b7c7-e6b96a0952e1", 00:16:02.949 "strip_size_kb": 0, 00:16:02.949 "state": "online", 00:16:02.949 "raid_level": "raid1", 00:16:02.949 "superblock": true, 00:16:02.949 "num_base_bdevs": 2, 00:16:02.949 "num_base_bdevs_discovered": 1, 00:16:02.949 "num_base_bdevs_operational": 1, 00:16:02.949 "base_bdevs_list": [ 00:16:02.949 { 00:16:02.949 "name": null, 00:16:02.949 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:02.949 "is_configured": false, 00:16:02.949 "data_offset": 0, 00:16:02.949 "data_size": 7936 00:16:02.949 }, 00:16:02.949 { 00:16:02.949 "name": "BaseBdev2", 00:16:02.949 "uuid": "13d2d49d-c27f-5472-8a1a-07c03d9e4fee", 00:16:02.949 "is_configured": true, 00:16:02.949 "data_offset": 256, 00:16:02.949 "data_size": 7936 00:16:02.949 } 00:16:02.949 ] 00:16:02.949 }' 00:16:02.949 15:20:31 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:02.949 15:20:31 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:03.527 15:20:31 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:03.527 15:20:31 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:03.527 15:20:31 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:03.527 [2024-11-27 15:20:31.388299] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:03.527 [2024-11-27 15:20:31.388529] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:16:03.527 [2024-11-27 15:20:31.388591] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:16:03.527 [2024-11-27 15:20:31.388649] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:03.527 [2024-11-27 15:20:31.393631] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c1a80 00:16:03.527 15:20:31 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:03.527 15:20:31 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@757 -- # sleep 1 00:16:03.527 [2024-11-27 15:20:31.395531] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:04.464 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:04.465 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:04.465 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:04.465 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:04.465 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:04.465 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:04.465 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.465 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:04.465 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:04.465 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.465 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:04.465 "name": "raid_bdev1", 00:16:04.465 "uuid": "3e22d42e-5cce-482d-b7c7-e6b96a0952e1", 00:16:04.465 "strip_size_kb": 0, 00:16:04.465 "state": "online", 00:16:04.465 "raid_level": "raid1", 00:16:04.465 "superblock": true, 00:16:04.465 "num_base_bdevs": 2, 00:16:04.465 "num_base_bdevs_discovered": 2, 00:16:04.465 "num_base_bdevs_operational": 2, 00:16:04.465 "process": { 00:16:04.465 "type": "rebuild", 00:16:04.465 "target": "spare", 00:16:04.465 "progress": { 00:16:04.465 "blocks": 2560, 00:16:04.465 "percent": 32 00:16:04.465 } 00:16:04.465 }, 00:16:04.465 "base_bdevs_list": [ 00:16:04.465 { 00:16:04.465 "name": "spare", 00:16:04.465 "uuid": "2a0b07f9-d30d-52b2-bf69-c956b4c05376", 00:16:04.465 "is_configured": true, 00:16:04.465 "data_offset": 256, 00:16:04.465 "data_size": 7936 00:16:04.465 }, 00:16:04.465 { 00:16:04.465 "name": "BaseBdev2", 00:16:04.465 "uuid": "13d2d49d-c27f-5472-8a1a-07c03d9e4fee", 00:16:04.465 "is_configured": true, 00:16:04.465 "data_offset": 256, 00:16:04.465 "data_size": 7936 00:16:04.465 } 00:16:04.465 ] 00:16:04.465 }' 00:16:04.465 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:04.465 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:04.465 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:04.465 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:04.465 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:16:04.465 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.465 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:04.465 [2024-11-27 15:20:32.563736] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:04.725 [2024-11-27 15:20:32.599597] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:04.725 [2024-11-27 15:20:32.599710] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:04.725 [2024-11-27 15:20:32.599728] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:04.725 [2024-11-27 15:20:32.599736] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:04.725 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.725 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:04.725 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:04.725 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:04.725 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:04.725 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:04.725 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:04.725 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:04.725 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:04.725 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:04.725 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:04.725 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:04.725 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.725 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:04.725 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:04.725 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.725 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:04.725 "name": "raid_bdev1", 00:16:04.725 "uuid": "3e22d42e-5cce-482d-b7c7-e6b96a0952e1", 00:16:04.725 "strip_size_kb": 0, 00:16:04.725 "state": "online", 00:16:04.725 "raid_level": "raid1", 00:16:04.725 "superblock": true, 00:16:04.725 "num_base_bdevs": 2, 00:16:04.725 "num_base_bdevs_discovered": 1, 00:16:04.725 "num_base_bdevs_operational": 1, 00:16:04.725 "base_bdevs_list": [ 00:16:04.725 { 00:16:04.725 "name": null, 00:16:04.725 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:04.725 "is_configured": false, 00:16:04.725 "data_offset": 0, 00:16:04.725 "data_size": 7936 00:16:04.725 }, 00:16:04.725 { 00:16:04.725 "name": "BaseBdev2", 00:16:04.725 "uuid": "13d2d49d-c27f-5472-8a1a-07c03d9e4fee", 00:16:04.725 "is_configured": true, 00:16:04.725 "data_offset": 256, 00:16:04.725 "data_size": 7936 00:16:04.725 } 00:16:04.725 ] 00:16:04.725 }' 00:16:04.725 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:04.725 15:20:32 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:04.985 15:20:33 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:04.985 15:20:33 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.985 15:20:33 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:04.985 [2024-11-27 15:20:33.087811] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:04.985 [2024-11-27 15:20:33.087913] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:04.985 [2024-11-27 15:20:33.087974] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:16:04.985 [2024-11-27 15:20:33.088046] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:04.985 [2024-11-27 15:20:33.088509] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:04.985 [2024-11-27 15:20:33.088568] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:04.985 [2024-11-27 15:20:33.088680] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:16:04.985 [2024-11-27 15:20:33.088720] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:16:04.985 [2024-11-27 15:20:33.088765] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:16:04.985 [2024-11-27 15:20:33.088805] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:05.245 [2024-11-27 15:20:33.093292] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c1b50 00:16:05.245 spare 00:16:05.245 15:20:33 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:05.245 15:20:33 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@764 -- # sleep 1 00:16:05.245 [2024-11-27 15:20:33.095202] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:06.184 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:06.184 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:06.184 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:06.184 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:06.184 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:06.184 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:06.184 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:06.184 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:06.184 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:06.184 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:06.184 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:06.184 "name": "raid_bdev1", 00:16:06.184 "uuid": "3e22d42e-5cce-482d-b7c7-e6b96a0952e1", 00:16:06.184 "strip_size_kb": 0, 00:16:06.184 "state": "online", 00:16:06.184 "raid_level": "raid1", 00:16:06.184 "superblock": true, 00:16:06.185 "num_base_bdevs": 2, 00:16:06.185 "num_base_bdevs_discovered": 2, 00:16:06.185 "num_base_bdevs_operational": 2, 00:16:06.185 "process": { 00:16:06.185 "type": "rebuild", 00:16:06.185 "target": "spare", 00:16:06.185 "progress": { 00:16:06.185 "blocks": 2560, 00:16:06.185 "percent": 32 00:16:06.185 } 00:16:06.185 }, 00:16:06.185 "base_bdevs_list": [ 00:16:06.185 { 00:16:06.185 "name": "spare", 00:16:06.185 "uuid": "2a0b07f9-d30d-52b2-bf69-c956b4c05376", 00:16:06.185 "is_configured": true, 00:16:06.185 "data_offset": 256, 00:16:06.185 "data_size": 7936 00:16:06.185 }, 00:16:06.185 { 00:16:06.185 "name": "BaseBdev2", 00:16:06.185 "uuid": "13d2d49d-c27f-5472-8a1a-07c03d9e4fee", 00:16:06.185 "is_configured": true, 00:16:06.185 "data_offset": 256, 00:16:06.185 "data_size": 7936 00:16:06.185 } 00:16:06.185 ] 00:16:06.185 }' 00:16:06.185 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:06.185 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:06.185 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:06.185 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:06.185 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:16:06.185 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:06.185 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:06.185 [2024-11-27 15:20:34.255393] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:06.445 [2024-11-27 15:20:34.299234] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:06.445 [2024-11-27 15:20:34.299288] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:06.445 [2024-11-27 15:20:34.299301] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:06.445 [2024-11-27 15:20:34.299309] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:06.445 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:06.445 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:06.445 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:06.445 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:06.445 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:06.445 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:06.445 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:06.445 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:06.445 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:06.445 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:06.445 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:06.445 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:06.445 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:06.445 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:06.445 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:06.445 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:06.445 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:06.445 "name": "raid_bdev1", 00:16:06.445 "uuid": "3e22d42e-5cce-482d-b7c7-e6b96a0952e1", 00:16:06.445 "strip_size_kb": 0, 00:16:06.445 "state": "online", 00:16:06.445 "raid_level": "raid1", 00:16:06.445 "superblock": true, 00:16:06.445 "num_base_bdevs": 2, 00:16:06.445 "num_base_bdevs_discovered": 1, 00:16:06.445 "num_base_bdevs_operational": 1, 00:16:06.445 "base_bdevs_list": [ 00:16:06.445 { 00:16:06.445 "name": null, 00:16:06.445 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:06.445 "is_configured": false, 00:16:06.445 "data_offset": 0, 00:16:06.445 "data_size": 7936 00:16:06.445 }, 00:16:06.445 { 00:16:06.445 "name": "BaseBdev2", 00:16:06.445 "uuid": "13d2d49d-c27f-5472-8a1a-07c03d9e4fee", 00:16:06.445 "is_configured": true, 00:16:06.445 "data_offset": 256, 00:16:06.445 "data_size": 7936 00:16:06.445 } 00:16:06.445 ] 00:16:06.445 }' 00:16:06.445 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:06.445 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:06.704 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:06.704 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:06.704 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:06.704 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:06.704 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:06.704 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:06.704 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:06.704 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:06.704 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:06.704 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:06.704 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:06.704 "name": "raid_bdev1", 00:16:06.704 "uuid": "3e22d42e-5cce-482d-b7c7-e6b96a0952e1", 00:16:06.704 "strip_size_kb": 0, 00:16:06.704 "state": "online", 00:16:06.704 "raid_level": "raid1", 00:16:06.704 "superblock": true, 00:16:06.704 "num_base_bdevs": 2, 00:16:06.704 "num_base_bdevs_discovered": 1, 00:16:06.704 "num_base_bdevs_operational": 1, 00:16:06.704 "base_bdevs_list": [ 00:16:06.704 { 00:16:06.704 "name": null, 00:16:06.704 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:06.704 "is_configured": false, 00:16:06.704 "data_offset": 0, 00:16:06.704 "data_size": 7936 00:16:06.704 }, 00:16:06.704 { 00:16:06.704 "name": "BaseBdev2", 00:16:06.704 "uuid": "13d2d49d-c27f-5472-8a1a-07c03d9e4fee", 00:16:06.704 "is_configured": true, 00:16:06.704 "data_offset": 256, 00:16:06.704 "data_size": 7936 00:16:06.704 } 00:16:06.704 ] 00:16:06.704 }' 00:16:06.704 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:06.964 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:06.964 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:06.964 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:06.964 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:16:06.964 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:06.964 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:06.964 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:06.964 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:16:06.964 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:06.964 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:06.964 [2024-11-27 15:20:34.878899] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:16:06.964 [2024-11-27 15:20:34.878962] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:06.964 [2024-11-27 15:20:34.878982] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:16:06.964 [2024-11-27 15:20:34.878993] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:06.964 [2024-11-27 15:20:34.879370] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:06.964 [2024-11-27 15:20:34.879388] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:06.964 [2024-11-27 15:20:34.879480] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:16:06.964 [2024-11-27 15:20:34.879498] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:16:06.964 [2024-11-27 15:20:34.879505] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:16:06.964 [2024-11-27 15:20:34.879520] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:16:06.964 BaseBdev1 00:16:06.964 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:06.964 15:20:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@775 -- # sleep 1 00:16:07.905 15:20:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:07.905 15:20:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:07.905 15:20:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:07.905 15:20:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:07.905 15:20:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:07.905 15:20:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:07.905 15:20:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:07.905 15:20:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:07.905 15:20:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:07.905 15:20:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:07.905 15:20:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:07.905 15:20:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:07.905 15:20:35 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:07.905 15:20:35 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:07.905 15:20:35 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:07.905 15:20:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:07.905 "name": "raid_bdev1", 00:16:07.905 "uuid": "3e22d42e-5cce-482d-b7c7-e6b96a0952e1", 00:16:07.905 "strip_size_kb": 0, 00:16:07.905 "state": "online", 00:16:07.905 "raid_level": "raid1", 00:16:07.905 "superblock": true, 00:16:07.905 "num_base_bdevs": 2, 00:16:07.905 "num_base_bdevs_discovered": 1, 00:16:07.905 "num_base_bdevs_operational": 1, 00:16:07.905 "base_bdevs_list": [ 00:16:07.905 { 00:16:07.905 "name": null, 00:16:07.905 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:07.905 "is_configured": false, 00:16:07.905 "data_offset": 0, 00:16:07.905 "data_size": 7936 00:16:07.905 }, 00:16:07.905 { 00:16:07.905 "name": "BaseBdev2", 00:16:07.905 "uuid": "13d2d49d-c27f-5472-8a1a-07c03d9e4fee", 00:16:07.905 "is_configured": true, 00:16:07.905 "data_offset": 256, 00:16:07.905 "data_size": 7936 00:16:07.905 } 00:16:07.905 ] 00:16:07.905 }' 00:16:07.905 15:20:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:07.905 15:20:35 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:08.475 15:20:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:08.475 15:20:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:08.475 15:20:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:08.475 15:20:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:08.475 15:20:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:08.475 15:20:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:08.475 15:20:36 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.475 15:20:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:08.475 15:20:36 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:08.475 15:20:36 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:08.475 15:20:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:08.475 "name": "raid_bdev1", 00:16:08.475 "uuid": "3e22d42e-5cce-482d-b7c7-e6b96a0952e1", 00:16:08.475 "strip_size_kb": 0, 00:16:08.475 "state": "online", 00:16:08.475 "raid_level": "raid1", 00:16:08.475 "superblock": true, 00:16:08.475 "num_base_bdevs": 2, 00:16:08.475 "num_base_bdevs_discovered": 1, 00:16:08.475 "num_base_bdevs_operational": 1, 00:16:08.475 "base_bdevs_list": [ 00:16:08.475 { 00:16:08.475 "name": null, 00:16:08.475 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:08.475 "is_configured": false, 00:16:08.475 "data_offset": 0, 00:16:08.475 "data_size": 7936 00:16:08.475 }, 00:16:08.475 { 00:16:08.475 "name": "BaseBdev2", 00:16:08.475 "uuid": "13d2d49d-c27f-5472-8a1a-07c03d9e4fee", 00:16:08.475 "is_configured": true, 00:16:08.475 "data_offset": 256, 00:16:08.475 "data_size": 7936 00:16:08.475 } 00:16:08.475 ] 00:16:08.475 }' 00:16:08.475 15:20:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:08.475 15:20:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:08.475 15:20:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:08.475 15:20:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:08.475 15:20:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:08.475 15:20:36 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@652 -- # local es=0 00:16:08.475 15:20:36 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:08.475 15:20:36 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:16:08.475 15:20:36 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:16:08.475 15:20:36 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:16:08.476 15:20:36 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:16:08.476 15:20:36 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:08.476 15:20:36 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.476 15:20:36 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:08.476 [2024-11-27 15:20:36.524051] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:08.476 [2024-11-27 15:20:36.524269] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:16:08.476 [2024-11-27 15:20:36.524285] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:16:08.476 request: 00:16:08.476 { 00:16:08.476 "base_bdev": "BaseBdev1", 00:16:08.476 "raid_bdev": "raid_bdev1", 00:16:08.476 "method": "bdev_raid_add_base_bdev", 00:16:08.476 "req_id": 1 00:16:08.476 } 00:16:08.476 Got JSON-RPC error response 00:16:08.476 response: 00:16:08.476 { 00:16:08.476 "code": -22, 00:16:08.476 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:16:08.476 } 00:16:08.476 15:20:36 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:16:08.476 15:20:36 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@655 -- # es=1 00:16:08.476 15:20:36 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:16:08.476 15:20:36 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:16:08.476 15:20:36 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:16:08.476 15:20:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@779 -- # sleep 1 00:16:09.854 15:20:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:09.854 15:20:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:09.854 15:20:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:09.854 15:20:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:09.854 15:20:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:09.854 15:20:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:09.854 15:20:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:09.854 15:20:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:09.854 15:20:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:09.854 15:20:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:09.854 15:20:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:09.854 15:20:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:09.854 15:20:37 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:09.854 15:20:37 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:09.854 15:20:37 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:09.854 15:20:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:09.854 "name": "raid_bdev1", 00:16:09.854 "uuid": "3e22d42e-5cce-482d-b7c7-e6b96a0952e1", 00:16:09.854 "strip_size_kb": 0, 00:16:09.854 "state": "online", 00:16:09.854 "raid_level": "raid1", 00:16:09.854 "superblock": true, 00:16:09.854 "num_base_bdevs": 2, 00:16:09.854 "num_base_bdevs_discovered": 1, 00:16:09.854 "num_base_bdevs_operational": 1, 00:16:09.854 "base_bdevs_list": [ 00:16:09.854 { 00:16:09.854 "name": null, 00:16:09.854 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:09.854 "is_configured": false, 00:16:09.854 "data_offset": 0, 00:16:09.854 "data_size": 7936 00:16:09.854 }, 00:16:09.854 { 00:16:09.854 "name": "BaseBdev2", 00:16:09.854 "uuid": "13d2d49d-c27f-5472-8a1a-07c03d9e4fee", 00:16:09.854 "is_configured": true, 00:16:09.854 "data_offset": 256, 00:16:09.854 "data_size": 7936 00:16:09.854 } 00:16:09.854 ] 00:16:09.854 }' 00:16:09.854 15:20:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:09.854 15:20:37 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:10.114 15:20:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:10.114 15:20:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:10.114 15:20:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:10.114 15:20:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:10.114 15:20:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:10.114 15:20:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:10.114 15:20:37 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:10.114 15:20:37 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:10.114 15:20:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:10.114 15:20:37 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:10.114 15:20:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:10.114 "name": "raid_bdev1", 00:16:10.114 "uuid": "3e22d42e-5cce-482d-b7c7-e6b96a0952e1", 00:16:10.114 "strip_size_kb": 0, 00:16:10.114 "state": "online", 00:16:10.114 "raid_level": "raid1", 00:16:10.114 "superblock": true, 00:16:10.114 "num_base_bdevs": 2, 00:16:10.114 "num_base_bdevs_discovered": 1, 00:16:10.114 "num_base_bdevs_operational": 1, 00:16:10.114 "base_bdevs_list": [ 00:16:10.114 { 00:16:10.114 "name": null, 00:16:10.114 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:10.114 "is_configured": false, 00:16:10.114 "data_offset": 0, 00:16:10.114 "data_size": 7936 00:16:10.114 }, 00:16:10.114 { 00:16:10.114 "name": "BaseBdev2", 00:16:10.114 "uuid": "13d2d49d-c27f-5472-8a1a-07c03d9e4fee", 00:16:10.114 "is_configured": true, 00:16:10.114 "data_offset": 256, 00:16:10.114 "data_size": 7936 00:16:10.114 } 00:16:10.114 ] 00:16:10.114 }' 00:16:10.114 15:20:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:10.114 15:20:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:10.114 15:20:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:10.114 15:20:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:10.114 15:20:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@784 -- # killprocess 97071 00:16:10.114 15:20:38 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@954 -- # '[' -z 97071 ']' 00:16:10.114 15:20:38 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@958 -- # kill -0 97071 00:16:10.114 15:20:38 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@959 -- # uname 00:16:10.114 15:20:38 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:10.114 15:20:38 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 97071 00:16:10.114 15:20:38 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:10.114 15:20:38 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:10.114 15:20:38 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@972 -- # echo 'killing process with pid 97071' 00:16:10.114 killing process with pid 97071 00:16:10.114 15:20:38 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@973 -- # kill 97071 00:16:10.114 Received shutdown signal, test time was about 60.000000 seconds 00:16:10.114 00:16:10.114 Latency(us) 00:16:10.114 [2024-11-27T15:20:38.221Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:16:10.114 [2024-11-27T15:20:38.221Z] =================================================================================================================== 00:16:10.114 [2024-11-27T15:20:38.221Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:16:10.114 [2024-11-27 15:20:38.179322] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:10.114 [2024-11-27 15:20:38.179446] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:10.114 [2024-11-27 15:20:38.179501] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:10.114 [2024-11-27 15:20:38.179510] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state offline 00:16:10.114 15:20:38 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@978 -- # wait 97071 00:16:10.114 [2024-11-27 15:20:38.212013] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:10.374 15:20:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@786 -- # return 0 00:16:10.374 00:16:10.374 real 0m18.269s 00:16:10.374 user 0m24.130s 00:16:10.374 sys 0m2.767s 00:16:10.374 ************************************ 00:16:10.374 END TEST raid_rebuild_test_sb_4k 00:16:10.374 ************************************ 00:16:10.374 15:20:38 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:10.374 15:20:38 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:10.634 15:20:38 bdev_raid -- bdev/bdev_raid.sh@1003 -- # base_malloc_params='-m 32' 00:16:10.634 15:20:38 bdev_raid -- bdev/bdev_raid.sh@1004 -- # run_test raid_state_function_test_sb_md_separate raid_state_function_test raid1 2 true 00:16:10.634 15:20:38 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:16:10.634 15:20:38 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:10.634 15:20:38 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:10.634 ************************************ 00:16:10.634 START TEST raid_state_function_test_sb_md_separate 00:16:10.634 ************************************ 00:16:10.634 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 true 00:16:10.634 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:16:10.634 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:16:10.635 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:16:10.635 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:16:10.635 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:16:10.635 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:10.635 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:16:10.635 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:16:10.635 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:10.635 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:16:10.635 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:16:10.635 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:10.635 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:16:10.635 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:16:10.635 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:16:10.635 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # local strip_size 00:16:10.635 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:16:10.635 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:16:10.635 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:16:10.635 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:16:10.635 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:16:10.635 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:16:10.635 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@229 -- # raid_pid=97745 00:16:10.635 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:16:10.635 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 97745' 00:16:10.635 Process raid pid: 97745 00:16:10.635 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@231 -- # waitforlisten 97745 00:16:10.635 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@835 -- # '[' -z 97745 ']' 00:16:10.635 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:10.635 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:10.635 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:10.635 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:10.635 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:10.635 15:20:38 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:10.635 [2024-11-27 15:20:38.609487] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:16:10.635 [2024-11-27 15:20:38.609709] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:16:10.895 [2024-11-27 15:20:38.793671] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:10.895 [2024-11-27 15:20:38.821714] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:10.895 [2024-11-27 15:20:38.866218] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:10.895 [2024-11-27 15:20:38.866264] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:11.463 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:11.463 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@868 -- # return 0 00:16:11.463 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:11.463 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:11.463 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:11.463 [2024-11-27 15:20:39.417731] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:16:11.463 [2024-11-27 15:20:39.417790] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:16:11.463 [2024-11-27 15:20:39.417800] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:11.463 [2024-11-27 15:20:39.417810] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:11.463 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:11.463 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:11.463 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:11.463 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:11.463 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:11.463 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:11.463 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:11.463 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:11.463 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:11.463 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:11.463 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:11.463 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:11.463 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:11.463 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:11.463 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:11.463 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:11.463 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:11.463 "name": "Existed_Raid", 00:16:11.463 "uuid": "a66a5fa9-c501-4055-aee5-f1d2c710da3a", 00:16:11.463 "strip_size_kb": 0, 00:16:11.463 "state": "configuring", 00:16:11.463 "raid_level": "raid1", 00:16:11.463 "superblock": true, 00:16:11.463 "num_base_bdevs": 2, 00:16:11.463 "num_base_bdevs_discovered": 0, 00:16:11.463 "num_base_bdevs_operational": 2, 00:16:11.463 "base_bdevs_list": [ 00:16:11.463 { 00:16:11.463 "name": "BaseBdev1", 00:16:11.463 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:11.463 "is_configured": false, 00:16:11.463 "data_offset": 0, 00:16:11.463 "data_size": 0 00:16:11.463 }, 00:16:11.463 { 00:16:11.463 "name": "BaseBdev2", 00:16:11.463 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:11.463 "is_configured": false, 00:16:11.463 "data_offset": 0, 00:16:11.463 "data_size": 0 00:16:11.463 } 00:16:11.463 ] 00:16:11.463 }' 00:16:11.463 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:11.463 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:12.032 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:16:12.032 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:12.032 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:12.032 [2024-11-27 15:20:39.892819] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:16:12.032 [2024-11-27 15:20:39.892911] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:16:12.032 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:12.032 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:12.032 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:12.032 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:12.032 [2024-11-27 15:20:39.905251] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:16:12.033 [2024-11-27 15:20:39.905567] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:16:12.033 [2024-11-27 15:20:39.905718] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:12.033 [2024-11-27 15:20:39.905868] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev1 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:12.033 [2024-11-27 15:20:39.938488] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:12.033 BaseBdev1 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@905 -- # local i 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:12.033 [ 00:16:12.033 { 00:16:12.033 "name": "BaseBdev1", 00:16:12.033 "aliases": [ 00:16:12.033 "c33ccb96-bcb4-4022-b6bc-457300e433a5" 00:16:12.033 ], 00:16:12.033 "product_name": "Malloc disk", 00:16:12.033 "block_size": 4096, 00:16:12.033 "num_blocks": 8192, 00:16:12.033 "uuid": "c33ccb96-bcb4-4022-b6bc-457300e433a5", 00:16:12.033 "md_size": 32, 00:16:12.033 "md_interleave": false, 00:16:12.033 "dif_type": 0, 00:16:12.033 "assigned_rate_limits": { 00:16:12.033 "rw_ios_per_sec": 0, 00:16:12.033 "rw_mbytes_per_sec": 0, 00:16:12.033 "r_mbytes_per_sec": 0, 00:16:12.033 "w_mbytes_per_sec": 0 00:16:12.033 }, 00:16:12.033 "claimed": true, 00:16:12.033 "claim_type": "exclusive_write", 00:16:12.033 "zoned": false, 00:16:12.033 "supported_io_types": { 00:16:12.033 "read": true, 00:16:12.033 "write": true, 00:16:12.033 "unmap": true, 00:16:12.033 "flush": true, 00:16:12.033 "reset": true, 00:16:12.033 "nvme_admin": false, 00:16:12.033 "nvme_io": false, 00:16:12.033 "nvme_io_md": false, 00:16:12.033 "write_zeroes": true, 00:16:12.033 "zcopy": true, 00:16:12.033 "get_zone_info": false, 00:16:12.033 "zone_management": false, 00:16:12.033 "zone_append": false, 00:16:12.033 "compare": false, 00:16:12.033 "compare_and_write": false, 00:16:12.033 "abort": true, 00:16:12.033 "seek_hole": false, 00:16:12.033 "seek_data": false, 00:16:12.033 "copy": true, 00:16:12.033 "nvme_iov_md": false 00:16:12.033 }, 00:16:12.033 "memory_domains": [ 00:16:12.033 { 00:16:12.033 "dma_device_id": "system", 00:16:12.033 "dma_device_type": 1 00:16:12.033 }, 00:16:12.033 { 00:16:12.033 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:12.033 "dma_device_type": 2 00:16:12.033 } 00:16:12.033 ], 00:16:12.033 "driver_specific": {} 00:16:12.033 } 00:16:12.033 ] 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@911 -- # return 0 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:12.033 15:20:39 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:12.033 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:12.033 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:12.033 "name": "Existed_Raid", 00:16:12.033 "uuid": "db098fdc-f5cf-4e87-927f-ed5c23d7d3bf", 00:16:12.033 "strip_size_kb": 0, 00:16:12.033 "state": "configuring", 00:16:12.033 "raid_level": "raid1", 00:16:12.033 "superblock": true, 00:16:12.033 "num_base_bdevs": 2, 00:16:12.033 "num_base_bdevs_discovered": 1, 00:16:12.033 "num_base_bdevs_operational": 2, 00:16:12.033 "base_bdevs_list": [ 00:16:12.033 { 00:16:12.033 "name": "BaseBdev1", 00:16:12.033 "uuid": "c33ccb96-bcb4-4022-b6bc-457300e433a5", 00:16:12.033 "is_configured": true, 00:16:12.033 "data_offset": 256, 00:16:12.033 "data_size": 7936 00:16:12.033 }, 00:16:12.033 { 00:16:12.033 "name": "BaseBdev2", 00:16:12.033 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:12.033 "is_configured": false, 00:16:12.033 "data_offset": 0, 00:16:12.033 "data_size": 0 00:16:12.033 } 00:16:12.033 ] 00:16:12.033 }' 00:16:12.033 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:12.033 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:12.603 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:16:12.603 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:12.603 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:12.603 [2024-11-27 15:20:40.413877] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:16:12.603 [2024-11-27 15:20:40.413939] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:16:12.603 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:12.603 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:12.603 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:12.603 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:12.603 [2024-11-27 15:20:40.421929] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:12.603 [2024-11-27 15:20:40.424027] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:12.603 [2024-11-27 15:20:40.424113] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:12.603 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:12.603 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:16:12.603 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:16:12.603 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:12.603 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:12.603 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:12.603 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:12.603 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:12.603 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:12.603 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:12.603 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:12.603 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:12.603 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:12.604 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:12.604 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:12.604 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:12.604 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:12.604 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:12.604 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:12.604 "name": "Existed_Raid", 00:16:12.604 "uuid": "aa5c8cc1-c14f-4e95-a073-d3702adf974a", 00:16:12.604 "strip_size_kb": 0, 00:16:12.604 "state": "configuring", 00:16:12.604 "raid_level": "raid1", 00:16:12.604 "superblock": true, 00:16:12.604 "num_base_bdevs": 2, 00:16:12.604 "num_base_bdevs_discovered": 1, 00:16:12.604 "num_base_bdevs_operational": 2, 00:16:12.604 "base_bdevs_list": [ 00:16:12.604 { 00:16:12.604 "name": "BaseBdev1", 00:16:12.604 "uuid": "c33ccb96-bcb4-4022-b6bc-457300e433a5", 00:16:12.604 "is_configured": true, 00:16:12.604 "data_offset": 256, 00:16:12.604 "data_size": 7936 00:16:12.604 }, 00:16:12.604 { 00:16:12.604 "name": "BaseBdev2", 00:16:12.604 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:12.604 "is_configured": false, 00:16:12.604 "data_offset": 0, 00:16:12.604 "data_size": 0 00:16:12.604 } 00:16:12.604 ] 00:16:12.604 }' 00:16:12.604 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:12.604 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:12.864 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev2 00:16:12.864 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:12.864 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:12.864 [2024-11-27 15:20:40.895183] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:12.864 [2024-11-27 15:20:40.895402] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:16:12.864 [2024-11-27 15:20:40.895427] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:12.864 [2024-11-27 15:20:40.895566] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:16:12.864 [2024-11-27 15:20:40.895708] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:16:12.864 [2024-11-27 15:20:40.895724] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:16:12.864 [2024-11-27 15:20:40.895824] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:12.864 BaseBdev2 00:16:12.864 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:12.864 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:16:12.864 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:16:12.864 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:16:12.864 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@905 -- # local i 00:16:12.864 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:16:12.864 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:16:12.864 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:16:12.864 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:12.864 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:12.864 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:12.864 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:16:12.864 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:12.864 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:12.864 [ 00:16:12.864 { 00:16:12.864 "name": "BaseBdev2", 00:16:12.864 "aliases": [ 00:16:12.864 "06b1a600-4854-4efe-b4c1-8a570237506b" 00:16:12.864 ], 00:16:12.864 "product_name": "Malloc disk", 00:16:12.864 "block_size": 4096, 00:16:12.864 "num_blocks": 8192, 00:16:12.864 "uuid": "06b1a600-4854-4efe-b4c1-8a570237506b", 00:16:12.864 "md_size": 32, 00:16:12.864 "md_interleave": false, 00:16:12.864 "dif_type": 0, 00:16:12.864 "assigned_rate_limits": { 00:16:12.864 "rw_ios_per_sec": 0, 00:16:12.864 "rw_mbytes_per_sec": 0, 00:16:12.864 "r_mbytes_per_sec": 0, 00:16:12.864 "w_mbytes_per_sec": 0 00:16:12.864 }, 00:16:12.864 "claimed": true, 00:16:12.864 "claim_type": "exclusive_write", 00:16:12.864 "zoned": false, 00:16:12.864 "supported_io_types": { 00:16:12.864 "read": true, 00:16:12.864 "write": true, 00:16:12.864 "unmap": true, 00:16:12.864 "flush": true, 00:16:12.864 "reset": true, 00:16:12.864 "nvme_admin": false, 00:16:12.864 "nvme_io": false, 00:16:12.864 "nvme_io_md": false, 00:16:12.864 "write_zeroes": true, 00:16:12.864 "zcopy": true, 00:16:12.864 "get_zone_info": false, 00:16:12.864 "zone_management": false, 00:16:12.864 "zone_append": false, 00:16:12.864 "compare": false, 00:16:12.864 "compare_and_write": false, 00:16:12.864 "abort": true, 00:16:12.864 "seek_hole": false, 00:16:12.864 "seek_data": false, 00:16:12.864 "copy": true, 00:16:12.864 "nvme_iov_md": false 00:16:12.864 }, 00:16:12.864 "memory_domains": [ 00:16:12.864 { 00:16:12.864 "dma_device_id": "system", 00:16:12.864 "dma_device_type": 1 00:16:12.864 }, 00:16:12.864 { 00:16:12.864 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:12.864 "dma_device_type": 2 00:16:12.864 } 00:16:12.864 ], 00:16:12.864 "driver_specific": {} 00:16:12.864 } 00:16:12.864 ] 00:16:12.864 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:12.864 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@911 -- # return 0 00:16:12.864 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:16:12.864 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:16:12.864 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:16:12.865 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:12.865 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:12.865 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:12.865 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:12.865 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:12.865 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:12.865 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:12.865 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:12.865 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:12.865 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:12.865 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:12.865 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:12.865 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:12.865 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:13.125 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:13.125 "name": "Existed_Raid", 00:16:13.125 "uuid": "aa5c8cc1-c14f-4e95-a073-d3702adf974a", 00:16:13.125 "strip_size_kb": 0, 00:16:13.125 "state": "online", 00:16:13.125 "raid_level": "raid1", 00:16:13.125 "superblock": true, 00:16:13.125 "num_base_bdevs": 2, 00:16:13.125 "num_base_bdevs_discovered": 2, 00:16:13.125 "num_base_bdevs_operational": 2, 00:16:13.125 "base_bdevs_list": [ 00:16:13.125 { 00:16:13.125 "name": "BaseBdev1", 00:16:13.126 "uuid": "c33ccb96-bcb4-4022-b6bc-457300e433a5", 00:16:13.126 "is_configured": true, 00:16:13.126 "data_offset": 256, 00:16:13.126 "data_size": 7936 00:16:13.126 }, 00:16:13.126 { 00:16:13.126 "name": "BaseBdev2", 00:16:13.126 "uuid": "06b1a600-4854-4efe-b4c1-8a570237506b", 00:16:13.126 "is_configured": true, 00:16:13.126 "data_offset": 256, 00:16:13.126 "data_size": 7936 00:16:13.126 } 00:16:13.126 ] 00:16:13.126 }' 00:16:13.126 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:13.126 15:20:40 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:13.386 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:16:13.386 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:16:13.386 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:13.386 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:13.386 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:16:13.386 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:13.386 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:13.386 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:16:13.386 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:13.386 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:13.386 [2024-11-27 15:20:41.358731] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:13.386 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:13.386 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:13.386 "name": "Existed_Raid", 00:16:13.386 "aliases": [ 00:16:13.386 "aa5c8cc1-c14f-4e95-a073-d3702adf974a" 00:16:13.386 ], 00:16:13.386 "product_name": "Raid Volume", 00:16:13.386 "block_size": 4096, 00:16:13.386 "num_blocks": 7936, 00:16:13.386 "uuid": "aa5c8cc1-c14f-4e95-a073-d3702adf974a", 00:16:13.386 "md_size": 32, 00:16:13.386 "md_interleave": false, 00:16:13.386 "dif_type": 0, 00:16:13.386 "assigned_rate_limits": { 00:16:13.386 "rw_ios_per_sec": 0, 00:16:13.386 "rw_mbytes_per_sec": 0, 00:16:13.386 "r_mbytes_per_sec": 0, 00:16:13.386 "w_mbytes_per_sec": 0 00:16:13.386 }, 00:16:13.386 "claimed": false, 00:16:13.386 "zoned": false, 00:16:13.386 "supported_io_types": { 00:16:13.386 "read": true, 00:16:13.386 "write": true, 00:16:13.386 "unmap": false, 00:16:13.386 "flush": false, 00:16:13.386 "reset": true, 00:16:13.386 "nvme_admin": false, 00:16:13.386 "nvme_io": false, 00:16:13.387 "nvme_io_md": false, 00:16:13.387 "write_zeroes": true, 00:16:13.387 "zcopy": false, 00:16:13.387 "get_zone_info": false, 00:16:13.387 "zone_management": false, 00:16:13.387 "zone_append": false, 00:16:13.387 "compare": false, 00:16:13.387 "compare_and_write": false, 00:16:13.387 "abort": false, 00:16:13.387 "seek_hole": false, 00:16:13.387 "seek_data": false, 00:16:13.387 "copy": false, 00:16:13.387 "nvme_iov_md": false 00:16:13.387 }, 00:16:13.387 "memory_domains": [ 00:16:13.387 { 00:16:13.387 "dma_device_id": "system", 00:16:13.387 "dma_device_type": 1 00:16:13.387 }, 00:16:13.387 { 00:16:13.387 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:13.387 "dma_device_type": 2 00:16:13.387 }, 00:16:13.387 { 00:16:13.387 "dma_device_id": "system", 00:16:13.387 "dma_device_type": 1 00:16:13.387 }, 00:16:13.387 { 00:16:13.387 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:13.387 "dma_device_type": 2 00:16:13.387 } 00:16:13.387 ], 00:16:13.387 "driver_specific": { 00:16:13.387 "raid": { 00:16:13.387 "uuid": "aa5c8cc1-c14f-4e95-a073-d3702adf974a", 00:16:13.387 "strip_size_kb": 0, 00:16:13.387 "state": "online", 00:16:13.387 "raid_level": "raid1", 00:16:13.387 "superblock": true, 00:16:13.387 "num_base_bdevs": 2, 00:16:13.387 "num_base_bdevs_discovered": 2, 00:16:13.387 "num_base_bdevs_operational": 2, 00:16:13.387 "base_bdevs_list": [ 00:16:13.387 { 00:16:13.387 "name": "BaseBdev1", 00:16:13.387 "uuid": "c33ccb96-bcb4-4022-b6bc-457300e433a5", 00:16:13.387 "is_configured": true, 00:16:13.387 "data_offset": 256, 00:16:13.387 "data_size": 7936 00:16:13.387 }, 00:16:13.387 { 00:16:13.387 "name": "BaseBdev2", 00:16:13.387 "uuid": "06b1a600-4854-4efe-b4c1-8a570237506b", 00:16:13.387 "is_configured": true, 00:16:13.387 "data_offset": 256, 00:16:13.387 "data_size": 7936 00:16:13.387 } 00:16:13.387 ] 00:16:13.387 } 00:16:13.387 } 00:16:13.387 }' 00:16:13.387 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:13.387 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:16:13.387 BaseBdev2' 00:16:13.387 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:13.387 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:16:13.387 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:13.387 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:16:13.387 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:13.387 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:13.387 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:13.646 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:13.646 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:16:13.646 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:16:13.646 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:13.646 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:16:13.646 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:13.646 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:13.646 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:13.646 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:13.646 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:16:13.646 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:16:13.646 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:16:13.646 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:13.646 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:13.646 [2024-11-27 15:20:41.586174] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:13.646 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:13.646 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@260 -- # local expected_state 00:16:13.646 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:16:13.646 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@198 -- # case $1 in 00:16:13.646 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@199 -- # return 0 00:16:13.646 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:16:13.646 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:16:13.647 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:13.647 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:13.647 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:13.647 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:13.647 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:13.647 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:13.647 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:13.647 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:13.647 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:13.647 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:13.647 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:13.647 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:13.647 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:13.647 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:13.647 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:13.647 "name": "Existed_Raid", 00:16:13.647 "uuid": "aa5c8cc1-c14f-4e95-a073-d3702adf974a", 00:16:13.647 "strip_size_kb": 0, 00:16:13.647 "state": "online", 00:16:13.647 "raid_level": "raid1", 00:16:13.647 "superblock": true, 00:16:13.647 "num_base_bdevs": 2, 00:16:13.647 "num_base_bdevs_discovered": 1, 00:16:13.647 "num_base_bdevs_operational": 1, 00:16:13.647 "base_bdevs_list": [ 00:16:13.647 { 00:16:13.647 "name": null, 00:16:13.647 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:13.647 "is_configured": false, 00:16:13.647 "data_offset": 0, 00:16:13.647 "data_size": 7936 00:16:13.647 }, 00:16:13.647 { 00:16:13.647 "name": "BaseBdev2", 00:16:13.647 "uuid": "06b1a600-4854-4efe-b4c1-8a570237506b", 00:16:13.647 "is_configured": true, 00:16:13.647 "data_offset": 256, 00:16:13.647 "data_size": 7936 00:16:13.647 } 00:16:13.647 ] 00:16:13.647 }' 00:16:13.647 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:13.647 15:20:41 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:14.216 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:16:14.216 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:16:14.216 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:14.216 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:16:14.216 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:14.216 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:14.216 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:14.216 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:16:14.216 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:16:14.216 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:16:14.216 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:14.216 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:14.216 [2024-11-27 15:20:42.091830] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:16:14.216 [2024-11-27 15:20:42.092023] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:14.216 [2024-11-27 15:20:42.114985] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:14.216 [2024-11-27 15:20:42.115117] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:14.216 [2024-11-27 15:20:42.115178] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:16:14.216 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:14.216 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:16:14.216 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:16:14.216 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:16:14.216 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:14.216 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:14.217 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:14.217 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:14.217 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:16:14.217 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:16:14.217 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:16:14.217 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@326 -- # killprocess 97745 00:16:14.217 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@954 -- # '[' -z 97745 ']' 00:16:14.217 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@958 -- # kill -0 97745 00:16:14.217 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@959 -- # uname 00:16:14.217 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:14.217 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 97745 00:16:14.217 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:14.217 killing process with pid 97745 00:16:14.217 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:14.217 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@972 -- # echo 'killing process with pid 97745' 00:16:14.217 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@973 -- # kill 97745 00:16:14.217 [2024-11-27 15:20:42.207238] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:14.217 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@978 -- # wait 97745 00:16:14.217 [2024-11-27 15:20:42.208846] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:14.477 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@328 -- # return 0 00:16:14.477 00:16:14.477 real 0m4.045s 00:16:14.477 user 0m6.227s 00:16:14.477 sys 0m0.886s 00:16:14.477 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:14.477 15:20:42 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:14.477 ************************************ 00:16:14.477 END TEST raid_state_function_test_sb_md_separate 00:16:14.477 ************************************ 00:16:14.737 15:20:42 bdev_raid -- bdev/bdev_raid.sh@1005 -- # run_test raid_superblock_test_md_separate raid_superblock_test raid1 2 00:16:14.738 15:20:42 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:16:14.738 15:20:42 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:14.738 15:20:42 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:14.738 ************************************ 00:16:14.738 START TEST raid_superblock_test_md_separate 00:16:14.738 ************************************ 00:16:14.738 15:20:42 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 2 00:16:14.738 15:20:42 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:16:14.738 15:20:42 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:16:14.738 15:20:42 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:16:14.738 15:20:42 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:16:14.738 15:20:42 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:16:14.738 15:20:42 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:16:14.738 15:20:42 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:16:14.738 15:20:42 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:16:14.738 15:20:42 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:16:14.738 15:20:42 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@399 -- # local strip_size 00:16:14.738 15:20:42 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:16:14.738 15:20:42 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:16:14.738 15:20:42 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:16:14.738 15:20:42 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:16:14.738 15:20:42 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:16:14.738 15:20:42 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@412 -- # raid_pid=97986 00:16:14.738 15:20:42 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:16:14.738 15:20:42 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@413 -- # waitforlisten 97986 00:16:14.738 15:20:42 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@835 -- # '[' -z 97986 ']' 00:16:14.738 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:14.738 15:20:42 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:14.738 15:20:42 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:14.738 15:20:42 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:14.738 15:20:42 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:14.738 15:20:42 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:14.738 [2024-11-27 15:20:42.717720] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:16:14.738 [2024-11-27 15:20:42.717907] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid97986 ] 00:16:14.998 [2024-11-27 15:20:42.889794] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:14.998 [2024-11-27 15:20:42.929960] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:14.998 [2024-11-27 15:20:43.006141] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:14.998 [2024-11-27 15:20:43.006286] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@868 -- # return 0 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b malloc1 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:15.569 malloc1 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:15.569 [2024-11-27 15:20:43.572179] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:15.569 [2024-11-27 15:20:43.572336] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:15.569 [2024-11-27 15:20:43.572380] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:16:15.569 [2024-11-27 15:20:43.572417] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:15.569 [2024-11-27 15:20:43.574611] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:15.569 [2024-11-27 15:20:43.574699] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:15.569 pt1 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b malloc2 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:15.569 malloc2 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:15.569 [2024-11-27 15:20:43.611965] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:15.569 [2024-11-27 15:20:43.612026] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:15.569 [2024-11-27 15:20:43.612046] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:16:15.569 [2024-11-27 15:20:43.612059] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:15.569 [2024-11-27 15:20:43.614186] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:15.569 [2024-11-27 15:20:43.614229] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:15.569 pt2 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:15.569 [2024-11-27 15:20:43.623958] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:15.569 [2024-11-27 15:20:43.625989] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:15.569 [2024-11-27 15:20:43.626159] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:16:15.569 [2024-11-27 15:20:43.626183] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:15.569 [2024-11-27 15:20:43.626289] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:16:15.569 [2024-11-27 15:20:43.626434] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:16:15.569 [2024-11-27 15:20:43.626445] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:16:15.569 [2024-11-27 15:20:43.626541] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:15.569 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:15.570 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:15.570 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:15.570 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:15.570 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:15.570 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:15.570 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:15.570 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:15.570 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:15.570 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:15.570 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:15.570 15:20:43 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:15.570 15:20:43 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:15.570 15:20:43 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:15.830 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:15.830 "name": "raid_bdev1", 00:16:15.830 "uuid": "63c458b2-bee2-4bbc-b208-21f064c9d178", 00:16:15.830 "strip_size_kb": 0, 00:16:15.830 "state": "online", 00:16:15.830 "raid_level": "raid1", 00:16:15.830 "superblock": true, 00:16:15.830 "num_base_bdevs": 2, 00:16:15.830 "num_base_bdevs_discovered": 2, 00:16:15.830 "num_base_bdevs_operational": 2, 00:16:15.830 "base_bdevs_list": [ 00:16:15.830 { 00:16:15.830 "name": "pt1", 00:16:15.830 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:15.830 "is_configured": true, 00:16:15.830 "data_offset": 256, 00:16:15.830 "data_size": 7936 00:16:15.830 }, 00:16:15.830 { 00:16:15.830 "name": "pt2", 00:16:15.830 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:15.830 "is_configured": true, 00:16:15.830 "data_offset": 256, 00:16:15.830 "data_size": 7936 00:16:15.830 } 00:16:15.830 ] 00:16:15.830 }' 00:16:15.830 15:20:43 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:15.830 15:20:43 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:16.090 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:16:16.090 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:16:16.090 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:16.090 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:16.090 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:16:16.090 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:16.090 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:16.090 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:16.090 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:16.090 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:16.090 [2024-11-27 15:20:44.087556] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:16.090 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:16.090 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:16.090 "name": "raid_bdev1", 00:16:16.090 "aliases": [ 00:16:16.090 "63c458b2-bee2-4bbc-b208-21f064c9d178" 00:16:16.090 ], 00:16:16.090 "product_name": "Raid Volume", 00:16:16.090 "block_size": 4096, 00:16:16.090 "num_blocks": 7936, 00:16:16.090 "uuid": "63c458b2-bee2-4bbc-b208-21f064c9d178", 00:16:16.090 "md_size": 32, 00:16:16.090 "md_interleave": false, 00:16:16.090 "dif_type": 0, 00:16:16.090 "assigned_rate_limits": { 00:16:16.090 "rw_ios_per_sec": 0, 00:16:16.090 "rw_mbytes_per_sec": 0, 00:16:16.090 "r_mbytes_per_sec": 0, 00:16:16.090 "w_mbytes_per_sec": 0 00:16:16.090 }, 00:16:16.090 "claimed": false, 00:16:16.090 "zoned": false, 00:16:16.090 "supported_io_types": { 00:16:16.090 "read": true, 00:16:16.090 "write": true, 00:16:16.090 "unmap": false, 00:16:16.090 "flush": false, 00:16:16.090 "reset": true, 00:16:16.090 "nvme_admin": false, 00:16:16.090 "nvme_io": false, 00:16:16.090 "nvme_io_md": false, 00:16:16.090 "write_zeroes": true, 00:16:16.090 "zcopy": false, 00:16:16.090 "get_zone_info": false, 00:16:16.090 "zone_management": false, 00:16:16.090 "zone_append": false, 00:16:16.090 "compare": false, 00:16:16.090 "compare_and_write": false, 00:16:16.090 "abort": false, 00:16:16.090 "seek_hole": false, 00:16:16.090 "seek_data": false, 00:16:16.090 "copy": false, 00:16:16.090 "nvme_iov_md": false 00:16:16.090 }, 00:16:16.090 "memory_domains": [ 00:16:16.090 { 00:16:16.090 "dma_device_id": "system", 00:16:16.090 "dma_device_type": 1 00:16:16.090 }, 00:16:16.090 { 00:16:16.090 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:16.090 "dma_device_type": 2 00:16:16.090 }, 00:16:16.090 { 00:16:16.090 "dma_device_id": "system", 00:16:16.090 "dma_device_type": 1 00:16:16.090 }, 00:16:16.090 { 00:16:16.090 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:16.090 "dma_device_type": 2 00:16:16.090 } 00:16:16.090 ], 00:16:16.090 "driver_specific": { 00:16:16.090 "raid": { 00:16:16.090 "uuid": "63c458b2-bee2-4bbc-b208-21f064c9d178", 00:16:16.090 "strip_size_kb": 0, 00:16:16.090 "state": "online", 00:16:16.090 "raid_level": "raid1", 00:16:16.090 "superblock": true, 00:16:16.090 "num_base_bdevs": 2, 00:16:16.090 "num_base_bdevs_discovered": 2, 00:16:16.090 "num_base_bdevs_operational": 2, 00:16:16.090 "base_bdevs_list": [ 00:16:16.090 { 00:16:16.090 "name": "pt1", 00:16:16.090 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:16.090 "is_configured": true, 00:16:16.090 "data_offset": 256, 00:16:16.090 "data_size": 7936 00:16:16.090 }, 00:16:16.090 { 00:16:16.090 "name": "pt2", 00:16:16.090 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:16.090 "is_configured": true, 00:16:16.090 "data_offset": 256, 00:16:16.090 "data_size": 7936 00:16:16.090 } 00:16:16.090 ] 00:16:16.090 } 00:16:16.090 } 00:16:16.090 }' 00:16:16.090 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:16.090 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:16:16.090 pt2' 00:16:16.090 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:16:16.351 [2024-11-27 15:20:44.303263] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=63c458b2-bee2-4bbc-b208-21f064c9d178 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@436 -- # '[' -z 63c458b2-bee2-4bbc-b208-21f064c9d178 ']' 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:16.351 [2024-11-27 15:20:44.351018] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:16.351 [2024-11-27 15:20:44.351046] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:16.351 [2024-11-27 15:20:44.351135] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:16.351 [2024-11-27 15:20:44.351199] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:16.351 [2024-11-27 15:20:44.351210] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:16.351 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@652 -- # local es=0 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:16.612 [2024-11-27 15:20:44.494960] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:16:16.612 [2024-11-27 15:20:44.497134] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:16:16.612 [2024-11-27 15:20:44.497243] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:16:16.612 [2024-11-27 15:20:44.497343] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:16:16.612 [2024-11-27 15:20:44.497405] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:16.612 [2024-11-27 15:20:44.497459] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state configuring 00:16:16.612 request: 00:16:16.612 { 00:16:16.612 "name": "raid_bdev1", 00:16:16.612 "raid_level": "raid1", 00:16:16.612 "base_bdevs": [ 00:16:16.612 "malloc1", 00:16:16.612 "malloc2" 00:16:16.612 ], 00:16:16.612 "superblock": false, 00:16:16.612 "method": "bdev_raid_create", 00:16:16.612 "req_id": 1 00:16:16.612 } 00:16:16.612 Got JSON-RPC error response 00:16:16.612 response: 00:16:16.612 { 00:16:16.612 "code": -17, 00:16:16.612 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:16:16.612 } 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@655 -- # es=1 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:16.612 [2024-11-27 15:20:44.562784] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:16.612 [2024-11-27 15:20:44.562879] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:16.612 [2024-11-27 15:20:44.562936] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:16:16.612 [2024-11-27 15:20:44.562970] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:16.612 [2024-11-27 15:20:44.565128] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:16.612 [2024-11-27 15:20:44.565202] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:16.612 [2024-11-27 15:20:44.565268] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:16:16.612 [2024-11-27 15:20:44.565333] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:16.612 pt1 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:16.612 "name": "raid_bdev1", 00:16:16.612 "uuid": "63c458b2-bee2-4bbc-b208-21f064c9d178", 00:16:16.612 "strip_size_kb": 0, 00:16:16.612 "state": "configuring", 00:16:16.612 "raid_level": "raid1", 00:16:16.612 "superblock": true, 00:16:16.612 "num_base_bdevs": 2, 00:16:16.612 "num_base_bdevs_discovered": 1, 00:16:16.612 "num_base_bdevs_operational": 2, 00:16:16.612 "base_bdevs_list": [ 00:16:16.612 { 00:16:16.612 "name": "pt1", 00:16:16.612 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:16.612 "is_configured": true, 00:16:16.612 "data_offset": 256, 00:16:16.612 "data_size": 7936 00:16:16.612 }, 00:16:16.612 { 00:16:16.612 "name": null, 00:16:16.612 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:16.612 "is_configured": false, 00:16:16.612 "data_offset": 256, 00:16:16.612 "data_size": 7936 00:16:16.612 } 00:16:16.612 ] 00:16:16.612 }' 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:16.612 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:16.874 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:16:16.874 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:16:16.874 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:16:16.875 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:16.875 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:16.875 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:16.875 [2024-11-27 15:20:44.962094] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:16.875 [2024-11-27 15:20:44.962205] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:16.875 [2024-11-27 15:20:44.962248] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:16:16.875 [2024-11-27 15:20:44.962281] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:16.875 [2024-11-27 15:20:44.962438] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:16.875 [2024-11-27 15:20:44.962491] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:16.875 [2024-11-27 15:20:44.962560] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:16:16.875 [2024-11-27 15:20:44.962606] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:16.875 [2024-11-27 15:20:44.962714] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:16:16.875 [2024-11-27 15:20:44.962753] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:16.875 [2024-11-27 15:20:44.962845] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:16:16.875 [2024-11-27 15:20:44.962973] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:16:16.875 [2024-11-27 15:20:44.963023] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:16:16.875 [2024-11-27 15:20:44.963120] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:16.875 pt2 00:16:16.875 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:16.875 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:16:16.875 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:16:16.875 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:16.875 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:16.875 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:16.875 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:16.875 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:16.875 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:16.875 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:16.875 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:16.875 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:16.875 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:16.875 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:16.875 15:20:44 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:16.875 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:16.875 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:17.135 15:20:44 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:17.135 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:17.135 "name": "raid_bdev1", 00:16:17.135 "uuid": "63c458b2-bee2-4bbc-b208-21f064c9d178", 00:16:17.135 "strip_size_kb": 0, 00:16:17.135 "state": "online", 00:16:17.135 "raid_level": "raid1", 00:16:17.135 "superblock": true, 00:16:17.135 "num_base_bdevs": 2, 00:16:17.135 "num_base_bdevs_discovered": 2, 00:16:17.135 "num_base_bdevs_operational": 2, 00:16:17.135 "base_bdevs_list": [ 00:16:17.135 { 00:16:17.135 "name": "pt1", 00:16:17.135 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:17.135 "is_configured": true, 00:16:17.135 "data_offset": 256, 00:16:17.135 "data_size": 7936 00:16:17.135 }, 00:16:17.135 { 00:16:17.135 "name": "pt2", 00:16:17.135 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:17.135 "is_configured": true, 00:16:17.135 "data_offset": 256, 00:16:17.135 "data_size": 7936 00:16:17.135 } 00:16:17.135 ] 00:16:17.135 }' 00:16:17.135 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:17.135 15:20:45 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:17.396 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:16:17.396 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:16:17.396 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:17.396 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:17.396 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:16:17.396 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:17.396 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:17.396 15:20:45 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:17.396 15:20:45 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:17.396 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:17.396 [2024-11-27 15:20:45.433531] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:17.396 15:20:45 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:17.396 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:17.396 "name": "raid_bdev1", 00:16:17.396 "aliases": [ 00:16:17.396 "63c458b2-bee2-4bbc-b208-21f064c9d178" 00:16:17.396 ], 00:16:17.396 "product_name": "Raid Volume", 00:16:17.396 "block_size": 4096, 00:16:17.396 "num_blocks": 7936, 00:16:17.396 "uuid": "63c458b2-bee2-4bbc-b208-21f064c9d178", 00:16:17.396 "md_size": 32, 00:16:17.396 "md_interleave": false, 00:16:17.396 "dif_type": 0, 00:16:17.396 "assigned_rate_limits": { 00:16:17.396 "rw_ios_per_sec": 0, 00:16:17.396 "rw_mbytes_per_sec": 0, 00:16:17.396 "r_mbytes_per_sec": 0, 00:16:17.396 "w_mbytes_per_sec": 0 00:16:17.396 }, 00:16:17.396 "claimed": false, 00:16:17.396 "zoned": false, 00:16:17.396 "supported_io_types": { 00:16:17.396 "read": true, 00:16:17.396 "write": true, 00:16:17.396 "unmap": false, 00:16:17.396 "flush": false, 00:16:17.396 "reset": true, 00:16:17.396 "nvme_admin": false, 00:16:17.396 "nvme_io": false, 00:16:17.396 "nvme_io_md": false, 00:16:17.396 "write_zeroes": true, 00:16:17.396 "zcopy": false, 00:16:17.396 "get_zone_info": false, 00:16:17.396 "zone_management": false, 00:16:17.396 "zone_append": false, 00:16:17.396 "compare": false, 00:16:17.396 "compare_and_write": false, 00:16:17.396 "abort": false, 00:16:17.396 "seek_hole": false, 00:16:17.396 "seek_data": false, 00:16:17.396 "copy": false, 00:16:17.396 "nvme_iov_md": false 00:16:17.396 }, 00:16:17.396 "memory_domains": [ 00:16:17.396 { 00:16:17.396 "dma_device_id": "system", 00:16:17.396 "dma_device_type": 1 00:16:17.396 }, 00:16:17.396 { 00:16:17.396 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:17.396 "dma_device_type": 2 00:16:17.396 }, 00:16:17.396 { 00:16:17.396 "dma_device_id": "system", 00:16:17.396 "dma_device_type": 1 00:16:17.396 }, 00:16:17.396 { 00:16:17.396 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:17.396 "dma_device_type": 2 00:16:17.396 } 00:16:17.396 ], 00:16:17.396 "driver_specific": { 00:16:17.396 "raid": { 00:16:17.396 "uuid": "63c458b2-bee2-4bbc-b208-21f064c9d178", 00:16:17.396 "strip_size_kb": 0, 00:16:17.396 "state": "online", 00:16:17.396 "raid_level": "raid1", 00:16:17.396 "superblock": true, 00:16:17.396 "num_base_bdevs": 2, 00:16:17.396 "num_base_bdevs_discovered": 2, 00:16:17.396 "num_base_bdevs_operational": 2, 00:16:17.396 "base_bdevs_list": [ 00:16:17.396 { 00:16:17.396 "name": "pt1", 00:16:17.396 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:17.396 "is_configured": true, 00:16:17.396 "data_offset": 256, 00:16:17.396 "data_size": 7936 00:16:17.396 }, 00:16:17.396 { 00:16:17.396 "name": "pt2", 00:16:17.396 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:17.396 "is_configured": true, 00:16:17.396 "data_offset": 256, 00:16:17.396 "data_size": 7936 00:16:17.396 } 00:16:17.396 ] 00:16:17.396 } 00:16:17.396 } 00:16:17.396 }' 00:16:17.396 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:17.657 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:16:17.657 pt2' 00:16:17.657 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:17.657 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:16:17.657 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:17.657 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:16:17.657 15:20:45 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:17.657 15:20:45 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:17.657 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:17.657 15:20:45 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:17.657 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:16:17.657 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:16:17.657 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:17.657 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:16:17.657 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:17.657 15:20:45 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:17.657 15:20:45 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:17.658 15:20:45 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:17.658 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:16:17.658 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:16:17.658 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:17.658 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:16:17.658 15:20:45 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:17.658 15:20:45 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:17.658 [2024-11-27 15:20:45.661204] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:17.658 15:20:45 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:17.658 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # '[' 63c458b2-bee2-4bbc-b208-21f064c9d178 '!=' 63c458b2-bee2-4bbc-b208-21f064c9d178 ']' 00:16:17.658 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:16:17.658 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@198 -- # case $1 in 00:16:17.658 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@199 -- # return 0 00:16:17.658 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:16:17.658 15:20:45 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:17.658 15:20:45 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:17.658 [2024-11-27 15:20:45.708910] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:16:17.658 15:20:45 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:17.658 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:17.658 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:17.658 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:17.658 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:17.658 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:17.658 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:17.658 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:17.658 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:17.658 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:17.658 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:17.658 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:17.658 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:17.658 15:20:45 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:17.658 15:20:45 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:17.658 15:20:45 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:17.937 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:17.937 "name": "raid_bdev1", 00:16:17.937 "uuid": "63c458b2-bee2-4bbc-b208-21f064c9d178", 00:16:17.937 "strip_size_kb": 0, 00:16:17.937 "state": "online", 00:16:17.937 "raid_level": "raid1", 00:16:17.937 "superblock": true, 00:16:17.937 "num_base_bdevs": 2, 00:16:17.937 "num_base_bdevs_discovered": 1, 00:16:17.937 "num_base_bdevs_operational": 1, 00:16:17.937 "base_bdevs_list": [ 00:16:17.937 { 00:16:17.937 "name": null, 00:16:17.937 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:17.937 "is_configured": false, 00:16:17.937 "data_offset": 0, 00:16:17.937 "data_size": 7936 00:16:17.937 }, 00:16:17.937 { 00:16:17.937 "name": "pt2", 00:16:17.937 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:17.937 "is_configured": true, 00:16:17.937 "data_offset": 256, 00:16:17.937 "data_size": 7936 00:16:17.937 } 00:16:17.937 ] 00:16:17.937 }' 00:16:17.937 15:20:45 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:17.937 15:20:45 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:18.210 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:18.210 15:20:46 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:18.210 15:20:46 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:18.210 [2024-11-27 15:20:46.192005] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:18.210 [2024-11-27 15:20:46.192087] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:18.210 [2024-11-27 15:20:46.192177] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:18.210 [2024-11-27 15:20:46.192239] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:18.210 [2024-11-27 15:20:46.192330] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:16:18.210 15:20:46 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:18.210 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:18.210 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:16:18.210 15:20:46 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:18.210 15:20:46 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:18.210 15:20:46 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:18.210 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:16:18.210 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:16:18.210 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:16:18.210 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:18.210 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:16:18.210 15:20:46 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:18.210 15:20:46 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:18.210 15:20:46 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:18.210 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:16:18.210 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:18.210 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:16:18.210 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:16:18.210 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@519 -- # i=1 00:16:18.210 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:18.210 15:20:46 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:18.210 15:20:46 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:18.210 [2024-11-27 15:20:46.268025] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:18.210 [2024-11-27 15:20:46.268117] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:18.210 [2024-11-27 15:20:46.268154] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:16:18.210 [2024-11-27 15:20:46.268185] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:18.210 [2024-11-27 15:20:46.270338] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:18.210 [2024-11-27 15:20:46.270418] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:18.210 [2024-11-27 15:20:46.270490] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:16:18.211 [2024-11-27 15:20:46.270538] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:18.211 [2024-11-27 15:20:46.270620] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:16:18.211 [2024-11-27 15:20:46.270685] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:18.211 [2024-11-27 15:20:46.270794] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:16:18.211 [2024-11-27 15:20:46.270936] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:16:18.211 [2024-11-27 15:20:46.270983] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006d00 00:16:18.211 [2024-11-27 15:20:46.271088] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:18.211 pt2 00:16:18.211 15:20:46 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:18.211 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:18.211 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:18.211 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:18.211 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:18.211 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:18.211 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:18.211 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:18.211 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:18.211 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:18.211 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:18.211 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:18.211 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:18.211 15:20:46 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:18.211 15:20:46 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:18.211 15:20:46 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:18.474 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:18.474 "name": "raid_bdev1", 00:16:18.474 "uuid": "63c458b2-bee2-4bbc-b208-21f064c9d178", 00:16:18.474 "strip_size_kb": 0, 00:16:18.474 "state": "online", 00:16:18.474 "raid_level": "raid1", 00:16:18.474 "superblock": true, 00:16:18.474 "num_base_bdevs": 2, 00:16:18.474 "num_base_bdevs_discovered": 1, 00:16:18.474 "num_base_bdevs_operational": 1, 00:16:18.474 "base_bdevs_list": [ 00:16:18.474 { 00:16:18.474 "name": null, 00:16:18.474 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:18.474 "is_configured": false, 00:16:18.474 "data_offset": 256, 00:16:18.474 "data_size": 7936 00:16:18.474 }, 00:16:18.474 { 00:16:18.474 "name": "pt2", 00:16:18.474 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:18.474 "is_configured": true, 00:16:18.474 "data_offset": 256, 00:16:18.474 "data_size": 7936 00:16:18.474 } 00:16:18.474 ] 00:16:18.474 }' 00:16:18.474 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:18.474 15:20:46 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:18.735 [2024-11-27 15:20:46.707496] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:18.735 [2024-11-27 15:20:46.707574] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:18.735 [2024-11-27 15:20:46.707636] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:18.735 [2024-11-27 15:20:46.707676] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:18.735 [2024-11-27 15:20:46.707689] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name raid_bdev1, state offline 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:18.735 [2024-11-27 15:20:46.767378] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:18.735 [2024-11-27 15:20:46.767445] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:18.735 [2024-11-27 15:20:46.767462] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:16:18.735 [2024-11-27 15:20:46.767478] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:18.735 [2024-11-27 15:20:46.769658] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:18.735 [2024-11-27 15:20:46.769704] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:18.735 [2024-11-27 15:20:46.769752] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:16:18.735 [2024-11-27 15:20:46.769795] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:18.735 [2024-11-27 15:20:46.769889] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:16:18.735 [2024-11-27 15:20:46.769925] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:18.735 [2024-11-27 15:20:46.769942] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007080 name raid_bdev1, state configuring 00:16:18.735 [2024-11-27 15:20:46.770003] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:18.735 [2024-11-27 15:20:46.770071] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007400 00:16:18.735 [2024-11-27 15:20:46.770082] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:18.735 [2024-11-27 15:20:46.770144] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:16:18.735 [2024-11-27 15:20:46.770235] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007400 00:16:18.735 [2024-11-27 15:20:46.770245] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007400 00:16:18.735 [2024-11-27 15:20:46.770325] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:18.735 pt1 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:18.735 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:18.735 "name": "raid_bdev1", 00:16:18.735 "uuid": "63c458b2-bee2-4bbc-b208-21f064c9d178", 00:16:18.735 "strip_size_kb": 0, 00:16:18.735 "state": "online", 00:16:18.735 "raid_level": "raid1", 00:16:18.735 "superblock": true, 00:16:18.735 "num_base_bdevs": 2, 00:16:18.735 "num_base_bdevs_discovered": 1, 00:16:18.735 "num_base_bdevs_operational": 1, 00:16:18.735 "base_bdevs_list": [ 00:16:18.735 { 00:16:18.735 "name": null, 00:16:18.735 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:18.736 "is_configured": false, 00:16:18.736 "data_offset": 256, 00:16:18.736 "data_size": 7936 00:16:18.736 }, 00:16:18.736 { 00:16:18.736 "name": "pt2", 00:16:18.736 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:18.736 "is_configured": true, 00:16:18.736 "data_offset": 256, 00:16:18.736 "data_size": 7936 00:16:18.736 } 00:16:18.736 ] 00:16:18.736 }' 00:16:18.736 15:20:46 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:18.736 15:20:46 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:19.306 15:20:47 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:16:19.306 15:20:47 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:19.306 15:20:47 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:19.306 15:20:47 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:16:19.306 15:20:47 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:19.306 15:20:47 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:16:19.306 15:20:47 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:19.306 15:20:47 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:19.306 15:20:47 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:16:19.306 15:20:47 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:19.306 [2024-11-27 15:20:47.270781] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:19.306 15:20:47 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:19.306 15:20:47 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # '[' 63c458b2-bee2-4bbc-b208-21f064c9d178 '!=' 63c458b2-bee2-4bbc-b208-21f064c9d178 ']' 00:16:19.306 15:20:47 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@563 -- # killprocess 97986 00:16:19.306 15:20:47 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@954 -- # '[' -z 97986 ']' 00:16:19.306 15:20:47 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@958 -- # kill -0 97986 00:16:19.306 15:20:47 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@959 -- # uname 00:16:19.306 15:20:47 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:19.306 15:20:47 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 97986 00:16:19.306 15:20:47 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:19.306 15:20:47 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:19.307 15:20:47 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@972 -- # echo 'killing process with pid 97986' 00:16:19.307 killing process with pid 97986 00:16:19.307 15:20:47 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@973 -- # kill 97986 00:16:19.307 [2024-11-27 15:20:47.347300] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:19.307 [2024-11-27 15:20:47.347371] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:19.307 [2024-11-27 15:20:47.347416] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:19.307 [2024-11-27 15:20:47.347436] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007400 name raid_bdev1, state offline 00:16:19.307 15:20:47 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@978 -- # wait 97986 00:16:19.307 [2024-11-27 15:20:47.393323] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:19.878 ************************************ 00:16:19.878 END TEST raid_superblock_test_md_separate 00:16:19.878 ************************************ 00:16:19.878 15:20:47 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@565 -- # return 0 00:16:19.878 00:16:19.878 real 0m5.110s 00:16:19.878 user 0m8.141s 00:16:19.878 sys 0m1.187s 00:16:19.878 15:20:47 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:19.878 15:20:47 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:19.878 15:20:47 bdev_raid -- bdev/bdev_raid.sh@1006 -- # '[' true = true ']' 00:16:19.878 15:20:47 bdev_raid -- bdev/bdev_raid.sh@1007 -- # run_test raid_rebuild_test_sb_md_separate raid_rebuild_test raid1 2 true false true 00:16:19.878 15:20:47 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:16:19.878 15:20:47 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:19.878 15:20:47 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:19.878 ************************************ 00:16:19.878 START TEST raid_rebuild_test_sb_md_separate 00:16:19.878 ************************************ 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true false true 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@573 -- # local verify=true 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # local strip_size 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@577 -- # local create_arg 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@579 -- # local data_offset 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@597 -- # raid_pid=98302 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@598 -- # waitforlisten 98302 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@835 -- # '[' -z 98302 ']' 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:19.878 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:19.878 15:20:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:19.878 [2024-11-27 15:20:47.922109] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:16:19.878 [2024-11-27 15:20:47.922325] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.ealI/O size of 3145728 is greater than zero copy threshold (65536). 00:16:19.878 Zero copy mechanism will not be used. 00:16:19.878 :6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid98302 ] 00:16:20.139 [2024-11-27 15:20:48.092269] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:20.139 [2024-11-27 15:20:48.133494] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:20.139 [2024-11-27 15:20:48.209503] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:20.139 [2024-11-27 15:20:48.209625] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:20.709 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:20.709 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@868 -- # return 0 00:16:20.709 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:20.709 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev1_malloc 00:16:20.709 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:20.709 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:20.709 BaseBdev1_malloc 00:16:20.709 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:20.709 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:16:20.709 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:20.709 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:20.709 [2024-11-27 15:20:48.763512] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:16:20.709 [2024-11-27 15:20:48.763678] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:20.709 [2024-11-27 15:20:48.763716] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:16:20.709 [2024-11-27 15:20:48.763740] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:20.709 [2024-11-27 15:20:48.765992] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:20.709 [2024-11-27 15:20:48.766031] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:20.709 BaseBdev1 00:16:20.709 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:20.709 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:20.709 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev2_malloc 00:16:20.709 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:20.709 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:20.709 BaseBdev2_malloc 00:16:20.710 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:20.710 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:16:20.710 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:20.710 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:20.710 [2024-11-27 15:20:48.799246] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:16:20.710 [2024-11-27 15:20:48.799304] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:20.710 [2024-11-27 15:20:48.799329] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:16:20.710 [2024-11-27 15:20:48.799339] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:20.710 [2024-11-27 15:20:48.801498] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:20.710 [2024-11-27 15:20:48.801537] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:16:20.710 BaseBdev2 00:16:20.710 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:20.710 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b spare_malloc 00:16:20.710 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:20.710 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:20.970 spare_malloc 00:16:20.970 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:20.970 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:16:20.970 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:20.970 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:20.970 spare_delay 00:16:20.970 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:20.970 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:20.970 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:20.970 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:20.970 [2024-11-27 15:20:48.864173] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:20.970 [2024-11-27 15:20:48.864273] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:20.970 [2024-11-27 15:20:48.864317] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:16:20.970 [2024-11-27 15:20:48.864337] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:20.970 [2024-11-27 15:20:48.867021] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:20.970 [2024-11-27 15:20:48.867067] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:20.970 spare 00:16:20.970 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:20.970 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:16:20.970 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:20.970 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:20.970 [2024-11-27 15:20:48.876137] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:20.970 [2024-11-27 15:20:48.878350] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:20.970 [2024-11-27 15:20:48.878630] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:16:20.970 [2024-11-27 15:20:48.878651] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:20.970 [2024-11-27 15:20:48.878742] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:16:20.970 [2024-11-27 15:20:48.878856] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:16:20.970 [2024-11-27 15:20:48.878874] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:16:20.970 [2024-11-27 15:20:48.878980] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:20.970 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:20.970 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:20.970 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:20.970 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:20.970 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:20.970 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:20.970 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:20.970 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:20.970 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:20.970 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:20.970 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:20.970 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:20.970 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:20.970 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:20.970 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:20.970 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:20.970 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:20.970 "name": "raid_bdev1", 00:16:20.970 "uuid": "27b5aa1d-72fc-452e-b440-88de9a7586b4", 00:16:20.970 "strip_size_kb": 0, 00:16:20.970 "state": "online", 00:16:20.970 "raid_level": "raid1", 00:16:20.970 "superblock": true, 00:16:20.970 "num_base_bdevs": 2, 00:16:20.970 "num_base_bdevs_discovered": 2, 00:16:20.970 "num_base_bdevs_operational": 2, 00:16:20.970 "base_bdevs_list": [ 00:16:20.970 { 00:16:20.970 "name": "BaseBdev1", 00:16:20.970 "uuid": "5aefcc68-3074-5064-bcd1-6a8da3580207", 00:16:20.970 "is_configured": true, 00:16:20.970 "data_offset": 256, 00:16:20.970 "data_size": 7936 00:16:20.970 }, 00:16:20.970 { 00:16:20.970 "name": "BaseBdev2", 00:16:20.971 "uuid": "1a122c50-ecfc-5f4e-aec7-6d485817f8d5", 00:16:20.971 "is_configured": true, 00:16:20.971 "data_offset": 256, 00:16:20.971 "data_size": 7936 00:16:20.971 } 00:16:20.971 ] 00:16:20.971 }' 00:16:20.971 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:20.971 15:20:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:21.232 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:21.232 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:21.232 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:21.232 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:16:21.232 [2024-11-27 15:20:49.331803] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:21.491 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:21.491 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:16:21.491 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:21.491 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:16:21.491 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:21.491 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:21.491 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:21.491 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:16:21.491 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:16:21.491 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:16:21.491 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:16:21.491 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:16:21.491 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:16:21.491 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:16:21.491 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # local bdev_list 00:16:21.491 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:16:21.491 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # local nbd_list 00:16:21.491 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@12 -- # local i 00:16:21.491 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:16:21.491 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:16:21.491 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:16:21.751 [2024-11-27 15:20:49.603452] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:16:21.751 /dev/nbd0 00:16:21.751 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:16:21.751 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:16:21.751 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:16:21.751 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # local i 00:16:21.751 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:16:21.751 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:16:21.751 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:16:21.751 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@877 -- # break 00:16:21.751 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:16:21.751 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:16:21.751 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:21.751 1+0 records in 00:16:21.751 1+0 records out 00:16:21.751 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000553998 s, 7.4 MB/s 00:16:21.751 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:21.751 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # size=4096 00:16:21.751 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:21.751 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:16:21.751 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@893 -- # return 0 00:16:21.751 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:21.751 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:16:21.751 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:16:21.751 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:16:21.751 15:20:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=4096 count=7936 oflag=direct 00:16:22.322 7936+0 records in 00:16:22.322 7936+0 records out 00:16:22.322 32505856 bytes (33 MB, 31 MiB) copied, 0.594252 s, 54.7 MB/s 00:16:22.322 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:16:22.322 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:16:22.322 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:16:22.322 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:22.322 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@51 -- # local i 00:16:22.322 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:22.322 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:16:22.583 [2024-11-27 15:20:50.471874] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:22.583 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:22.583 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:22.583 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:22.583 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:22.583 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:22.583 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:22.583 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:16:22.583 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:16:22.583 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:16:22.583 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:22.583 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:22.583 [2024-11-27 15:20:50.499898] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:22.583 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:22.583 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:22.583 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:22.583 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:22.583 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:22.583 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:22.583 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:22.583 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:22.583 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:22.583 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:22.583 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:22.583 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:22.583 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:22.583 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:22.583 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:22.583 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:22.583 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:22.583 "name": "raid_bdev1", 00:16:22.583 "uuid": "27b5aa1d-72fc-452e-b440-88de9a7586b4", 00:16:22.583 "strip_size_kb": 0, 00:16:22.583 "state": "online", 00:16:22.583 "raid_level": "raid1", 00:16:22.583 "superblock": true, 00:16:22.583 "num_base_bdevs": 2, 00:16:22.583 "num_base_bdevs_discovered": 1, 00:16:22.583 "num_base_bdevs_operational": 1, 00:16:22.583 "base_bdevs_list": [ 00:16:22.583 { 00:16:22.583 "name": null, 00:16:22.583 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:22.583 "is_configured": false, 00:16:22.583 "data_offset": 0, 00:16:22.583 "data_size": 7936 00:16:22.583 }, 00:16:22.583 { 00:16:22.583 "name": "BaseBdev2", 00:16:22.583 "uuid": "1a122c50-ecfc-5f4e-aec7-6d485817f8d5", 00:16:22.583 "is_configured": true, 00:16:22.583 "data_offset": 256, 00:16:22.583 "data_size": 7936 00:16:22.583 } 00:16:22.583 ] 00:16:22.583 }' 00:16:22.583 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:22.583 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:23.154 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:23.154 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.154 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:23.154 [2024-11-27 15:20:50.951188] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:23.154 [2024-11-27 15:20:50.953950] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00018d0c0 00:16:23.154 [2024-11-27 15:20:50.955829] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:23.154 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.154 15:20:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@647 -- # sleep 1 00:16:24.093 15:20:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:24.093 15:20:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:24.093 15:20:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:24.093 15:20:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:24.093 15:20:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:24.093 15:20:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:24.093 15:20:51 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:24.093 15:20:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:24.093 15:20:51 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:24.093 15:20:51 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:24.093 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:24.093 "name": "raid_bdev1", 00:16:24.093 "uuid": "27b5aa1d-72fc-452e-b440-88de9a7586b4", 00:16:24.093 "strip_size_kb": 0, 00:16:24.093 "state": "online", 00:16:24.093 "raid_level": "raid1", 00:16:24.093 "superblock": true, 00:16:24.093 "num_base_bdevs": 2, 00:16:24.093 "num_base_bdevs_discovered": 2, 00:16:24.093 "num_base_bdevs_operational": 2, 00:16:24.093 "process": { 00:16:24.093 "type": "rebuild", 00:16:24.093 "target": "spare", 00:16:24.093 "progress": { 00:16:24.093 "blocks": 2560, 00:16:24.093 "percent": 32 00:16:24.093 } 00:16:24.093 }, 00:16:24.093 "base_bdevs_list": [ 00:16:24.093 { 00:16:24.093 "name": "spare", 00:16:24.093 "uuid": "a14f9ba0-0341-574e-8685-cfd27fc42e0d", 00:16:24.093 "is_configured": true, 00:16:24.093 "data_offset": 256, 00:16:24.093 "data_size": 7936 00:16:24.093 }, 00:16:24.093 { 00:16:24.093 "name": "BaseBdev2", 00:16:24.093 "uuid": "1a122c50-ecfc-5f4e-aec7-6d485817f8d5", 00:16:24.093 "is_configured": true, 00:16:24.093 "data_offset": 256, 00:16:24.093 "data_size": 7936 00:16:24.093 } 00:16:24.093 ] 00:16:24.093 }' 00:16:24.093 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:24.093 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:24.093 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:24.093 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:24.093 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:24.093 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:24.093 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:24.093 [2024-11-27 15:20:52.122584] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:24.094 [2024-11-27 15:20:52.160600] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:24.094 [2024-11-27 15:20:52.160735] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:24.094 [2024-11-27 15:20:52.160758] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:24.094 [2024-11-27 15:20:52.160766] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:24.094 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:24.094 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:24.094 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:24.094 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:24.094 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:24.094 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:24.094 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:24.094 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:24.094 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:24.094 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:24.094 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:24.094 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:24.094 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:24.094 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:24.094 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:24.094 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:24.353 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:24.353 "name": "raid_bdev1", 00:16:24.353 "uuid": "27b5aa1d-72fc-452e-b440-88de9a7586b4", 00:16:24.353 "strip_size_kb": 0, 00:16:24.353 "state": "online", 00:16:24.353 "raid_level": "raid1", 00:16:24.353 "superblock": true, 00:16:24.353 "num_base_bdevs": 2, 00:16:24.353 "num_base_bdevs_discovered": 1, 00:16:24.353 "num_base_bdevs_operational": 1, 00:16:24.353 "base_bdevs_list": [ 00:16:24.353 { 00:16:24.353 "name": null, 00:16:24.353 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:24.353 "is_configured": false, 00:16:24.353 "data_offset": 0, 00:16:24.353 "data_size": 7936 00:16:24.353 }, 00:16:24.353 { 00:16:24.353 "name": "BaseBdev2", 00:16:24.353 "uuid": "1a122c50-ecfc-5f4e-aec7-6d485817f8d5", 00:16:24.353 "is_configured": true, 00:16:24.353 "data_offset": 256, 00:16:24.353 "data_size": 7936 00:16:24.353 } 00:16:24.353 ] 00:16:24.353 }' 00:16:24.353 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:24.354 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:24.613 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:24.613 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:24.613 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:24.613 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:24.613 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:24.613 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:24.613 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:24.613 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:24.613 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:24.613 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:24.613 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:24.613 "name": "raid_bdev1", 00:16:24.613 "uuid": "27b5aa1d-72fc-452e-b440-88de9a7586b4", 00:16:24.613 "strip_size_kb": 0, 00:16:24.613 "state": "online", 00:16:24.613 "raid_level": "raid1", 00:16:24.613 "superblock": true, 00:16:24.613 "num_base_bdevs": 2, 00:16:24.613 "num_base_bdevs_discovered": 1, 00:16:24.613 "num_base_bdevs_operational": 1, 00:16:24.613 "base_bdevs_list": [ 00:16:24.613 { 00:16:24.613 "name": null, 00:16:24.613 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:24.613 "is_configured": false, 00:16:24.613 "data_offset": 0, 00:16:24.614 "data_size": 7936 00:16:24.614 }, 00:16:24.614 { 00:16:24.614 "name": "BaseBdev2", 00:16:24.614 "uuid": "1a122c50-ecfc-5f4e-aec7-6d485817f8d5", 00:16:24.614 "is_configured": true, 00:16:24.614 "data_offset": 256, 00:16:24.614 "data_size": 7936 00:16:24.614 } 00:16:24.614 ] 00:16:24.614 }' 00:16:24.614 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:24.873 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:24.873 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:24.873 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:24.873 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:24.873 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:24.873 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:24.873 [2024-11-27 15:20:52.806929] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:24.873 [2024-11-27 15:20:52.809601] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00018d190 00:16:24.873 [2024-11-27 15:20:52.811566] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:24.873 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:24.873 15:20:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@663 -- # sleep 1 00:16:25.814 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:25.814 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:25.814 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:25.814 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:25.814 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:25.814 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:25.814 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:25.814 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:25.814 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:25.814 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:25.814 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:25.814 "name": "raid_bdev1", 00:16:25.814 "uuid": "27b5aa1d-72fc-452e-b440-88de9a7586b4", 00:16:25.814 "strip_size_kb": 0, 00:16:25.814 "state": "online", 00:16:25.814 "raid_level": "raid1", 00:16:25.814 "superblock": true, 00:16:25.814 "num_base_bdevs": 2, 00:16:25.814 "num_base_bdevs_discovered": 2, 00:16:25.814 "num_base_bdevs_operational": 2, 00:16:25.814 "process": { 00:16:25.814 "type": "rebuild", 00:16:25.814 "target": "spare", 00:16:25.814 "progress": { 00:16:25.814 "blocks": 2560, 00:16:25.814 "percent": 32 00:16:25.814 } 00:16:25.814 }, 00:16:25.814 "base_bdevs_list": [ 00:16:25.814 { 00:16:25.814 "name": "spare", 00:16:25.814 "uuid": "a14f9ba0-0341-574e-8685-cfd27fc42e0d", 00:16:25.814 "is_configured": true, 00:16:25.814 "data_offset": 256, 00:16:25.814 "data_size": 7936 00:16:25.814 }, 00:16:25.814 { 00:16:25.814 "name": "BaseBdev2", 00:16:25.814 "uuid": "1a122c50-ecfc-5f4e-aec7-6d485817f8d5", 00:16:25.814 "is_configured": true, 00:16:25.814 "data_offset": 256, 00:16:25.814 "data_size": 7936 00:16:25.814 } 00:16:25.814 ] 00:16:25.814 }' 00:16:25.814 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:25.814 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:25.814 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:26.074 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:26.074 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:16:26.074 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:16:26.074 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:16:26.074 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:16:26.074 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:16:26.074 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:16:26.074 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@706 -- # local timeout=594 00:16:26.074 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:26.074 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:26.074 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:26.074 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:26.074 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:26.074 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:26.074 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:26.074 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:26.074 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:26.074 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:26.074 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:26.074 15:20:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:26.074 "name": "raid_bdev1", 00:16:26.074 "uuid": "27b5aa1d-72fc-452e-b440-88de9a7586b4", 00:16:26.074 "strip_size_kb": 0, 00:16:26.074 "state": "online", 00:16:26.074 "raid_level": "raid1", 00:16:26.074 "superblock": true, 00:16:26.074 "num_base_bdevs": 2, 00:16:26.074 "num_base_bdevs_discovered": 2, 00:16:26.074 "num_base_bdevs_operational": 2, 00:16:26.074 "process": { 00:16:26.074 "type": "rebuild", 00:16:26.074 "target": "spare", 00:16:26.074 "progress": { 00:16:26.074 "blocks": 2816, 00:16:26.074 "percent": 35 00:16:26.074 } 00:16:26.074 }, 00:16:26.074 "base_bdevs_list": [ 00:16:26.074 { 00:16:26.074 "name": "spare", 00:16:26.074 "uuid": "a14f9ba0-0341-574e-8685-cfd27fc42e0d", 00:16:26.074 "is_configured": true, 00:16:26.074 "data_offset": 256, 00:16:26.074 "data_size": 7936 00:16:26.074 }, 00:16:26.074 { 00:16:26.074 "name": "BaseBdev2", 00:16:26.074 "uuid": "1a122c50-ecfc-5f4e-aec7-6d485817f8d5", 00:16:26.074 "is_configured": true, 00:16:26.074 "data_offset": 256, 00:16:26.074 "data_size": 7936 00:16:26.074 } 00:16:26.074 ] 00:16:26.074 }' 00:16:26.074 15:20:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:26.074 15:20:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:26.074 15:20:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:26.074 15:20:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:26.074 15:20:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:27.014 15:20:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:27.014 15:20:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:27.014 15:20:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:27.014 15:20:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:27.014 15:20:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:27.014 15:20:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:27.014 15:20:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:27.014 15:20:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:27.014 15:20:55 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:27.014 15:20:55 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:27.014 15:20:55 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:27.275 15:20:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:27.275 "name": "raid_bdev1", 00:16:27.275 "uuid": "27b5aa1d-72fc-452e-b440-88de9a7586b4", 00:16:27.275 "strip_size_kb": 0, 00:16:27.275 "state": "online", 00:16:27.275 "raid_level": "raid1", 00:16:27.275 "superblock": true, 00:16:27.275 "num_base_bdevs": 2, 00:16:27.275 "num_base_bdevs_discovered": 2, 00:16:27.275 "num_base_bdevs_operational": 2, 00:16:27.275 "process": { 00:16:27.275 "type": "rebuild", 00:16:27.275 "target": "spare", 00:16:27.275 "progress": { 00:16:27.275 "blocks": 5632, 00:16:27.275 "percent": 70 00:16:27.275 } 00:16:27.275 }, 00:16:27.275 "base_bdevs_list": [ 00:16:27.275 { 00:16:27.275 "name": "spare", 00:16:27.275 "uuid": "a14f9ba0-0341-574e-8685-cfd27fc42e0d", 00:16:27.275 "is_configured": true, 00:16:27.275 "data_offset": 256, 00:16:27.275 "data_size": 7936 00:16:27.275 }, 00:16:27.275 { 00:16:27.275 "name": "BaseBdev2", 00:16:27.275 "uuid": "1a122c50-ecfc-5f4e-aec7-6d485817f8d5", 00:16:27.275 "is_configured": true, 00:16:27.275 "data_offset": 256, 00:16:27.275 "data_size": 7936 00:16:27.275 } 00:16:27.275 ] 00:16:27.275 }' 00:16:27.275 15:20:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:27.275 15:20:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:27.275 15:20:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:27.275 15:20:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:27.275 15:20:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:27.850 [2024-11-27 15:20:55.922275] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:16:27.850 [2024-11-27 15:20:55.922401] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:16:27.850 [2024-11-27 15:20:55.922507] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:28.421 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:28.421 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:28.421 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:28.421 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:28.421 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:28.421 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:28.421 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:28.421 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:28.421 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:28.421 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:28.421 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:28.421 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:28.421 "name": "raid_bdev1", 00:16:28.421 "uuid": "27b5aa1d-72fc-452e-b440-88de9a7586b4", 00:16:28.421 "strip_size_kb": 0, 00:16:28.421 "state": "online", 00:16:28.421 "raid_level": "raid1", 00:16:28.421 "superblock": true, 00:16:28.421 "num_base_bdevs": 2, 00:16:28.421 "num_base_bdevs_discovered": 2, 00:16:28.421 "num_base_bdevs_operational": 2, 00:16:28.421 "base_bdevs_list": [ 00:16:28.421 { 00:16:28.421 "name": "spare", 00:16:28.421 "uuid": "a14f9ba0-0341-574e-8685-cfd27fc42e0d", 00:16:28.421 "is_configured": true, 00:16:28.421 "data_offset": 256, 00:16:28.421 "data_size": 7936 00:16:28.421 }, 00:16:28.421 { 00:16:28.421 "name": "BaseBdev2", 00:16:28.421 "uuid": "1a122c50-ecfc-5f4e-aec7-6d485817f8d5", 00:16:28.421 "is_configured": true, 00:16:28.421 "data_offset": 256, 00:16:28.421 "data_size": 7936 00:16:28.421 } 00:16:28.421 ] 00:16:28.421 }' 00:16:28.421 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:28.421 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:16:28.421 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:28.421 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:16:28.421 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@709 -- # break 00:16:28.421 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:28.421 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:28.421 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:28.421 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:28.421 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:28.421 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:28.421 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:28.421 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:28.421 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:28.421 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:28.421 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:28.421 "name": "raid_bdev1", 00:16:28.421 "uuid": "27b5aa1d-72fc-452e-b440-88de9a7586b4", 00:16:28.421 "strip_size_kb": 0, 00:16:28.421 "state": "online", 00:16:28.421 "raid_level": "raid1", 00:16:28.421 "superblock": true, 00:16:28.421 "num_base_bdevs": 2, 00:16:28.421 "num_base_bdevs_discovered": 2, 00:16:28.421 "num_base_bdevs_operational": 2, 00:16:28.421 "base_bdevs_list": [ 00:16:28.421 { 00:16:28.421 "name": "spare", 00:16:28.421 "uuid": "a14f9ba0-0341-574e-8685-cfd27fc42e0d", 00:16:28.421 "is_configured": true, 00:16:28.421 "data_offset": 256, 00:16:28.421 "data_size": 7936 00:16:28.421 }, 00:16:28.421 { 00:16:28.421 "name": "BaseBdev2", 00:16:28.421 "uuid": "1a122c50-ecfc-5f4e-aec7-6d485817f8d5", 00:16:28.421 "is_configured": true, 00:16:28.421 "data_offset": 256, 00:16:28.421 "data_size": 7936 00:16:28.421 } 00:16:28.421 ] 00:16:28.421 }' 00:16:28.421 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:28.421 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:28.421 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:28.682 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:28.682 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:28.682 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:28.682 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:28.682 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:28.682 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:28.682 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:28.682 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:28.682 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:28.682 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:28.682 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:28.682 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:28.682 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:28.682 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:28.682 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:28.682 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:28.682 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:28.682 "name": "raid_bdev1", 00:16:28.682 "uuid": "27b5aa1d-72fc-452e-b440-88de9a7586b4", 00:16:28.682 "strip_size_kb": 0, 00:16:28.682 "state": "online", 00:16:28.682 "raid_level": "raid1", 00:16:28.682 "superblock": true, 00:16:28.682 "num_base_bdevs": 2, 00:16:28.682 "num_base_bdevs_discovered": 2, 00:16:28.682 "num_base_bdevs_operational": 2, 00:16:28.682 "base_bdevs_list": [ 00:16:28.682 { 00:16:28.682 "name": "spare", 00:16:28.682 "uuid": "a14f9ba0-0341-574e-8685-cfd27fc42e0d", 00:16:28.682 "is_configured": true, 00:16:28.682 "data_offset": 256, 00:16:28.682 "data_size": 7936 00:16:28.682 }, 00:16:28.682 { 00:16:28.682 "name": "BaseBdev2", 00:16:28.682 "uuid": "1a122c50-ecfc-5f4e-aec7-6d485817f8d5", 00:16:28.682 "is_configured": true, 00:16:28.682 "data_offset": 256, 00:16:28.682 "data_size": 7936 00:16:28.682 } 00:16:28.682 ] 00:16:28.682 }' 00:16:28.682 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:28.682 15:20:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:28.942 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:28.942 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:28.942 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:28.942 [2024-11-27 15:20:57.031746] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:28.942 [2024-11-27 15:20:57.031815] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:28.942 [2024-11-27 15:20:57.031951] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:28.942 [2024-11-27 15:20:57.032032] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:28.942 [2024-11-27 15:20:57.032083] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:16:28.942 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:28.942 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:28.942 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # jq length 00:16:28.942 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:28.942 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:29.202 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:29.203 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:16:29.203 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:16:29.203 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:16:29.203 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:16:29.203 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:16:29.203 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:16:29.203 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # local bdev_list 00:16:29.203 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:16:29.203 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # local nbd_list 00:16:29.203 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@12 -- # local i 00:16:29.203 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:16:29.203 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:29.203 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:16:29.203 /dev/nbd0 00:16:29.462 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:16:29.463 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:16:29.463 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:16:29.463 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # local i 00:16:29.463 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:16:29.463 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:16:29.463 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:16:29.463 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@877 -- # break 00:16:29.463 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:16:29.463 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:16:29.463 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:29.463 1+0 records in 00:16:29.463 1+0 records out 00:16:29.463 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000380489 s, 10.8 MB/s 00:16:29.463 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:29.463 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # size=4096 00:16:29.463 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:29.463 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:16:29.463 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@893 -- # return 0 00:16:29.463 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:29.463 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:29.463 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:16:29.463 /dev/nbd1 00:16:29.463 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:16:29.723 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:16:29.723 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:16:29.723 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # local i 00:16:29.723 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:16:29.723 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:16:29.723 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:16:29.723 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@877 -- # break 00:16:29.723 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:16:29.723 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:16:29.723 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:29.723 1+0 records in 00:16:29.723 1+0 records out 00:16:29.723 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000371505 s, 11.0 MB/s 00:16:29.723 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:29.723 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # size=4096 00:16:29.723 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:29.723 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:16:29.723 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@893 -- # return 0 00:16:29.723 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:29.723 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:29.723 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:16:29.723 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:16:29.723 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:16:29.723 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:16:29.723 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:29.723 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@51 -- # local i 00:16:29.723 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:29.723 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:16:29.983 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:29.983 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:29.983 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:29.983 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:29.983 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:29.983 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:29.983 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:16:29.983 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:16:29.983 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:29.983 15:20:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:30.243 [2024-11-27 15:20:58.119889] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:30.243 [2024-11-27 15:20:58.119954] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:30.243 [2024-11-27 15:20:58.119976] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:16:30.243 [2024-11-27 15:20:58.119988] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:30.243 [2024-11-27 15:20:58.121981] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:30.243 [2024-11-27 15:20:58.122022] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:30.243 [2024-11-27 15:20:58.122075] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:16:30.243 [2024-11-27 15:20:58.122136] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:30.243 [2024-11-27 15:20:58.122266] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:30.243 spare 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:30.243 [2024-11-27 15:20:58.222167] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006600 00:16:30.243 [2024-11-27 15:20:58.222192] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:30.243 [2024-11-27 15:20:58.222313] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c19b0 00:16:30.243 [2024-11-27 15:20:58.222416] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006600 00:16:30.243 [2024-11-27 15:20:58.222427] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006600 00:16:30.243 [2024-11-27 15:20:58.222518] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:30.243 "name": "raid_bdev1", 00:16:30.243 "uuid": "27b5aa1d-72fc-452e-b440-88de9a7586b4", 00:16:30.243 "strip_size_kb": 0, 00:16:30.243 "state": "online", 00:16:30.243 "raid_level": "raid1", 00:16:30.243 "superblock": true, 00:16:30.243 "num_base_bdevs": 2, 00:16:30.243 "num_base_bdevs_discovered": 2, 00:16:30.243 "num_base_bdevs_operational": 2, 00:16:30.243 "base_bdevs_list": [ 00:16:30.243 { 00:16:30.243 "name": "spare", 00:16:30.243 "uuid": "a14f9ba0-0341-574e-8685-cfd27fc42e0d", 00:16:30.243 "is_configured": true, 00:16:30.243 "data_offset": 256, 00:16:30.243 "data_size": 7936 00:16:30.243 }, 00:16:30.243 { 00:16:30.243 "name": "BaseBdev2", 00:16:30.243 "uuid": "1a122c50-ecfc-5f4e-aec7-6d485817f8d5", 00:16:30.243 "is_configured": true, 00:16:30.243 "data_offset": 256, 00:16:30.243 "data_size": 7936 00:16:30.243 } 00:16:30.243 ] 00:16:30.243 }' 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:30.243 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:30.813 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:30.813 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:30.813 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:30.813 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:30.813 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:30.813 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:30.813 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:30.813 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:30.813 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:30.813 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:30.813 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:30.813 "name": "raid_bdev1", 00:16:30.813 "uuid": "27b5aa1d-72fc-452e-b440-88de9a7586b4", 00:16:30.813 "strip_size_kb": 0, 00:16:30.813 "state": "online", 00:16:30.813 "raid_level": "raid1", 00:16:30.813 "superblock": true, 00:16:30.813 "num_base_bdevs": 2, 00:16:30.813 "num_base_bdevs_discovered": 2, 00:16:30.813 "num_base_bdevs_operational": 2, 00:16:30.813 "base_bdevs_list": [ 00:16:30.813 { 00:16:30.813 "name": "spare", 00:16:30.813 "uuid": "a14f9ba0-0341-574e-8685-cfd27fc42e0d", 00:16:30.813 "is_configured": true, 00:16:30.813 "data_offset": 256, 00:16:30.813 "data_size": 7936 00:16:30.813 }, 00:16:30.813 { 00:16:30.813 "name": "BaseBdev2", 00:16:30.813 "uuid": "1a122c50-ecfc-5f4e-aec7-6d485817f8d5", 00:16:30.813 "is_configured": true, 00:16:30.813 "data_offset": 256, 00:16:30.813 "data_size": 7936 00:16:30.813 } 00:16:30.813 ] 00:16:30.813 }' 00:16:30.813 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:30.813 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:30.813 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:30.813 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:30.813 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:30.814 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:30.814 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:30.814 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:16:30.814 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:30.814 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:16:30.814 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:30.814 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:30.814 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:30.814 [2024-11-27 15:20:58.870638] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:30.814 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:30.814 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:30.814 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:30.814 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:30.814 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:30.814 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:30.814 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:30.814 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:30.814 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:30.814 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:30.814 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:30.814 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:30.814 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:30.814 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:30.814 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:30.814 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:31.073 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:31.074 "name": "raid_bdev1", 00:16:31.074 "uuid": "27b5aa1d-72fc-452e-b440-88de9a7586b4", 00:16:31.074 "strip_size_kb": 0, 00:16:31.074 "state": "online", 00:16:31.074 "raid_level": "raid1", 00:16:31.074 "superblock": true, 00:16:31.074 "num_base_bdevs": 2, 00:16:31.074 "num_base_bdevs_discovered": 1, 00:16:31.074 "num_base_bdevs_operational": 1, 00:16:31.074 "base_bdevs_list": [ 00:16:31.074 { 00:16:31.074 "name": null, 00:16:31.074 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:31.074 "is_configured": false, 00:16:31.074 "data_offset": 0, 00:16:31.074 "data_size": 7936 00:16:31.074 }, 00:16:31.074 { 00:16:31.074 "name": "BaseBdev2", 00:16:31.074 "uuid": "1a122c50-ecfc-5f4e-aec7-6d485817f8d5", 00:16:31.074 "is_configured": true, 00:16:31.074 "data_offset": 256, 00:16:31.074 "data_size": 7936 00:16:31.074 } 00:16:31.074 ] 00:16:31.074 }' 00:16:31.074 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:31.074 15:20:58 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:31.332 15:20:59 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:31.332 15:20:59 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:31.332 15:20:59 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:31.332 [2024-11-27 15:20:59.329880] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:31.332 [2024-11-27 15:20:59.330057] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:16:31.332 [2024-11-27 15:20:59.330088] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:16:31.332 [2024-11-27 15:20:59.330149] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:31.332 [2024-11-27 15:20:59.332653] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c1a80 00:16:31.332 [2024-11-27 15:20:59.334471] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:31.333 15:20:59 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:31.333 15:20:59 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@757 -- # sleep 1 00:16:32.272 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:32.272 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:32.272 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:32.272 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:32.272 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:32.272 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:32.272 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:32.272 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:32.272 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:32.272 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:32.533 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:32.533 "name": "raid_bdev1", 00:16:32.533 "uuid": "27b5aa1d-72fc-452e-b440-88de9a7586b4", 00:16:32.533 "strip_size_kb": 0, 00:16:32.533 "state": "online", 00:16:32.533 "raid_level": "raid1", 00:16:32.533 "superblock": true, 00:16:32.533 "num_base_bdevs": 2, 00:16:32.533 "num_base_bdevs_discovered": 2, 00:16:32.533 "num_base_bdevs_operational": 2, 00:16:32.533 "process": { 00:16:32.533 "type": "rebuild", 00:16:32.533 "target": "spare", 00:16:32.533 "progress": { 00:16:32.533 "blocks": 2560, 00:16:32.533 "percent": 32 00:16:32.533 } 00:16:32.533 }, 00:16:32.533 "base_bdevs_list": [ 00:16:32.533 { 00:16:32.533 "name": "spare", 00:16:32.533 "uuid": "a14f9ba0-0341-574e-8685-cfd27fc42e0d", 00:16:32.533 "is_configured": true, 00:16:32.533 "data_offset": 256, 00:16:32.533 "data_size": 7936 00:16:32.533 }, 00:16:32.533 { 00:16:32.533 "name": "BaseBdev2", 00:16:32.533 "uuid": "1a122c50-ecfc-5f4e-aec7-6d485817f8d5", 00:16:32.533 "is_configured": true, 00:16:32.533 "data_offset": 256, 00:16:32.533 "data_size": 7936 00:16:32.533 } 00:16:32.533 ] 00:16:32.533 }' 00:16:32.533 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:32.533 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:32.533 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:32.533 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:32.534 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:16:32.534 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:32.534 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:32.534 [2024-11-27 15:21:00.497316] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:32.534 [2024-11-27 15:21:00.538589] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:32.534 [2024-11-27 15:21:00.538642] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:32.534 [2024-11-27 15:21:00.538672] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:32.534 [2024-11-27 15:21:00.538679] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:32.534 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:32.534 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:32.534 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:32.534 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:32.534 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:32.534 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:32.534 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:32.534 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:32.534 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:32.534 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:32.534 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:32.534 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:32.534 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:32.534 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:32.534 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:32.534 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:32.534 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:32.534 "name": "raid_bdev1", 00:16:32.534 "uuid": "27b5aa1d-72fc-452e-b440-88de9a7586b4", 00:16:32.534 "strip_size_kb": 0, 00:16:32.534 "state": "online", 00:16:32.534 "raid_level": "raid1", 00:16:32.534 "superblock": true, 00:16:32.534 "num_base_bdevs": 2, 00:16:32.534 "num_base_bdevs_discovered": 1, 00:16:32.534 "num_base_bdevs_operational": 1, 00:16:32.534 "base_bdevs_list": [ 00:16:32.534 { 00:16:32.534 "name": null, 00:16:32.534 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:32.534 "is_configured": false, 00:16:32.534 "data_offset": 0, 00:16:32.534 "data_size": 7936 00:16:32.534 }, 00:16:32.534 { 00:16:32.534 "name": "BaseBdev2", 00:16:32.534 "uuid": "1a122c50-ecfc-5f4e-aec7-6d485817f8d5", 00:16:32.534 "is_configured": true, 00:16:32.534 "data_offset": 256, 00:16:32.534 "data_size": 7936 00:16:32.534 } 00:16:32.534 ] 00:16:32.534 }' 00:16:32.534 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:32.534 15:21:00 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:33.104 15:21:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:33.104 15:21:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:33.104 15:21:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:33.104 [2024-11-27 15:21:01.032875] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:33.104 [2024-11-27 15:21:01.032956] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:33.104 [2024-11-27 15:21:01.032983] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:16:33.104 [2024-11-27 15:21:01.032992] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:33.104 [2024-11-27 15:21:01.033209] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:33.104 [2024-11-27 15:21:01.033230] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:33.104 [2024-11-27 15:21:01.033284] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:16:33.104 [2024-11-27 15:21:01.033296] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:16:33.104 [2024-11-27 15:21:01.033320] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:16:33.104 [2024-11-27 15:21:01.033340] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:33.104 [2024-11-27 15:21:01.035478] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c1b50 00:16:33.104 [2024-11-27 15:21:01.037341] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:33.104 spare 00:16:33.104 15:21:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:33.104 15:21:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@764 -- # sleep 1 00:16:34.046 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:34.046 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:34.046 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:34.046 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:34.046 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:34.046 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:34.046 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:34.046 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:34.046 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:34.046 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:34.046 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:34.046 "name": "raid_bdev1", 00:16:34.046 "uuid": "27b5aa1d-72fc-452e-b440-88de9a7586b4", 00:16:34.046 "strip_size_kb": 0, 00:16:34.046 "state": "online", 00:16:34.046 "raid_level": "raid1", 00:16:34.046 "superblock": true, 00:16:34.046 "num_base_bdevs": 2, 00:16:34.046 "num_base_bdevs_discovered": 2, 00:16:34.046 "num_base_bdevs_operational": 2, 00:16:34.046 "process": { 00:16:34.046 "type": "rebuild", 00:16:34.046 "target": "spare", 00:16:34.046 "progress": { 00:16:34.046 "blocks": 2560, 00:16:34.046 "percent": 32 00:16:34.046 } 00:16:34.046 }, 00:16:34.046 "base_bdevs_list": [ 00:16:34.046 { 00:16:34.046 "name": "spare", 00:16:34.046 "uuid": "a14f9ba0-0341-574e-8685-cfd27fc42e0d", 00:16:34.046 "is_configured": true, 00:16:34.046 "data_offset": 256, 00:16:34.046 "data_size": 7936 00:16:34.046 }, 00:16:34.046 { 00:16:34.046 "name": "BaseBdev2", 00:16:34.046 "uuid": "1a122c50-ecfc-5f4e-aec7-6d485817f8d5", 00:16:34.046 "is_configured": true, 00:16:34.046 "data_offset": 256, 00:16:34.046 "data_size": 7936 00:16:34.046 } 00:16:34.046 ] 00:16:34.046 }' 00:16:34.046 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:34.046 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:34.046 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:34.307 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:34.307 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:16:34.307 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:34.307 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:34.307 [2024-11-27 15:21:02.180266] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:34.307 [2024-11-27 15:21:02.241529] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:34.307 [2024-11-27 15:21:02.241588] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:34.307 [2024-11-27 15:21:02.241602] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:34.307 [2024-11-27 15:21:02.241610] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:34.307 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:34.307 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:34.307 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:34.307 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:34.307 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:34.307 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:34.307 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:34.307 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:34.307 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:34.307 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:34.307 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:34.307 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:34.307 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:34.307 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:34.307 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:34.307 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:34.307 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:34.307 "name": "raid_bdev1", 00:16:34.307 "uuid": "27b5aa1d-72fc-452e-b440-88de9a7586b4", 00:16:34.307 "strip_size_kb": 0, 00:16:34.307 "state": "online", 00:16:34.307 "raid_level": "raid1", 00:16:34.307 "superblock": true, 00:16:34.307 "num_base_bdevs": 2, 00:16:34.307 "num_base_bdevs_discovered": 1, 00:16:34.307 "num_base_bdevs_operational": 1, 00:16:34.307 "base_bdevs_list": [ 00:16:34.307 { 00:16:34.307 "name": null, 00:16:34.307 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:34.307 "is_configured": false, 00:16:34.307 "data_offset": 0, 00:16:34.307 "data_size": 7936 00:16:34.307 }, 00:16:34.307 { 00:16:34.308 "name": "BaseBdev2", 00:16:34.308 "uuid": "1a122c50-ecfc-5f4e-aec7-6d485817f8d5", 00:16:34.308 "is_configured": true, 00:16:34.308 "data_offset": 256, 00:16:34.308 "data_size": 7936 00:16:34.308 } 00:16:34.308 ] 00:16:34.308 }' 00:16:34.308 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:34.308 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:34.879 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:34.879 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:34.879 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:34.879 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:34.879 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:34.879 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:34.879 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:34.879 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:34.879 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:34.879 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:34.879 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:34.879 "name": "raid_bdev1", 00:16:34.879 "uuid": "27b5aa1d-72fc-452e-b440-88de9a7586b4", 00:16:34.879 "strip_size_kb": 0, 00:16:34.879 "state": "online", 00:16:34.879 "raid_level": "raid1", 00:16:34.879 "superblock": true, 00:16:34.879 "num_base_bdevs": 2, 00:16:34.879 "num_base_bdevs_discovered": 1, 00:16:34.879 "num_base_bdevs_operational": 1, 00:16:34.879 "base_bdevs_list": [ 00:16:34.879 { 00:16:34.879 "name": null, 00:16:34.879 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:34.879 "is_configured": false, 00:16:34.879 "data_offset": 0, 00:16:34.879 "data_size": 7936 00:16:34.879 }, 00:16:34.879 { 00:16:34.879 "name": "BaseBdev2", 00:16:34.879 "uuid": "1a122c50-ecfc-5f4e-aec7-6d485817f8d5", 00:16:34.879 "is_configured": true, 00:16:34.879 "data_offset": 256, 00:16:34.879 "data_size": 7936 00:16:34.879 } 00:16:34.879 ] 00:16:34.879 }' 00:16:34.879 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:34.879 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:34.879 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:34.879 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:34.879 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:16:34.879 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:34.879 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:34.879 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:34.879 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:16:34.879 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:34.879 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:34.879 [2024-11-27 15:21:02.867536] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:16:34.879 [2024-11-27 15:21:02.867591] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:34.879 [2024-11-27 15:21:02.867614] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:16:34.879 [2024-11-27 15:21:02.867624] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:34.879 [2024-11-27 15:21:02.867835] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:34.879 [2024-11-27 15:21:02.867857] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:34.879 [2024-11-27 15:21:02.867918] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:16:34.879 [2024-11-27 15:21:02.867935] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:16:34.879 [2024-11-27 15:21:02.867943] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:16:34.879 [2024-11-27 15:21:02.867954] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:16:34.879 BaseBdev1 00:16:34.879 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:34.879 15:21:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@775 -- # sleep 1 00:16:35.819 15:21:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:35.819 15:21:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:35.819 15:21:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:35.819 15:21:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:35.819 15:21:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:35.819 15:21:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:35.819 15:21:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:35.819 15:21:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:35.819 15:21:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:35.819 15:21:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:35.819 15:21:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:35.819 15:21:03 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:35.819 15:21:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:35.819 15:21:03 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:35.819 15:21:03 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:36.079 15:21:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:36.079 "name": "raid_bdev1", 00:16:36.079 "uuid": "27b5aa1d-72fc-452e-b440-88de9a7586b4", 00:16:36.079 "strip_size_kb": 0, 00:16:36.079 "state": "online", 00:16:36.079 "raid_level": "raid1", 00:16:36.079 "superblock": true, 00:16:36.079 "num_base_bdevs": 2, 00:16:36.079 "num_base_bdevs_discovered": 1, 00:16:36.079 "num_base_bdevs_operational": 1, 00:16:36.079 "base_bdevs_list": [ 00:16:36.079 { 00:16:36.079 "name": null, 00:16:36.079 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:36.079 "is_configured": false, 00:16:36.079 "data_offset": 0, 00:16:36.079 "data_size": 7936 00:16:36.079 }, 00:16:36.079 { 00:16:36.079 "name": "BaseBdev2", 00:16:36.079 "uuid": "1a122c50-ecfc-5f4e-aec7-6d485817f8d5", 00:16:36.079 "is_configured": true, 00:16:36.079 "data_offset": 256, 00:16:36.079 "data_size": 7936 00:16:36.079 } 00:16:36.079 ] 00:16:36.079 }' 00:16:36.079 15:21:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:36.079 15:21:03 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:36.339 15:21:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:36.339 15:21:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:36.339 15:21:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:36.339 15:21:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:36.339 15:21:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:36.339 15:21:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:36.339 15:21:04 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:36.339 15:21:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:36.339 15:21:04 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:36.339 15:21:04 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:36.339 15:21:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:36.339 "name": "raid_bdev1", 00:16:36.339 "uuid": "27b5aa1d-72fc-452e-b440-88de9a7586b4", 00:16:36.339 "strip_size_kb": 0, 00:16:36.339 "state": "online", 00:16:36.339 "raid_level": "raid1", 00:16:36.339 "superblock": true, 00:16:36.339 "num_base_bdevs": 2, 00:16:36.339 "num_base_bdevs_discovered": 1, 00:16:36.339 "num_base_bdevs_operational": 1, 00:16:36.339 "base_bdevs_list": [ 00:16:36.339 { 00:16:36.339 "name": null, 00:16:36.339 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:36.339 "is_configured": false, 00:16:36.339 "data_offset": 0, 00:16:36.339 "data_size": 7936 00:16:36.339 }, 00:16:36.339 { 00:16:36.339 "name": "BaseBdev2", 00:16:36.339 "uuid": "1a122c50-ecfc-5f4e-aec7-6d485817f8d5", 00:16:36.339 "is_configured": true, 00:16:36.339 "data_offset": 256, 00:16:36.339 "data_size": 7936 00:16:36.339 } 00:16:36.339 ] 00:16:36.339 }' 00:16:36.339 15:21:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:36.339 15:21:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:36.339 15:21:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:36.599 15:21:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:36.599 15:21:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:36.599 15:21:04 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@652 -- # local es=0 00:16:36.599 15:21:04 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:36.599 15:21:04 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:16:36.599 15:21:04 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:16:36.599 15:21:04 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:16:36.599 15:21:04 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:16:36.599 15:21:04 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:36.599 15:21:04 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:36.599 15:21:04 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:36.599 [2024-11-27 15:21:04.460854] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:36.599 [2024-11-27 15:21:04.461038] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:16:36.599 [2024-11-27 15:21:04.461068] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:16:36.599 request: 00:16:36.599 { 00:16:36.599 "base_bdev": "BaseBdev1", 00:16:36.599 "raid_bdev": "raid_bdev1", 00:16:36.599 "method": "bdev_raid_add_base_bdev", 00:16:36.599 "req_id": 1 00:16:36.599 } 00:16:36.599 Got JSON-RPC error response 00:16:36.599 response: 00:16:36.599 { 00:16:36.599 "code": -22, 00:16:36.599 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:16:36.599 } 00:16:36.599 15:21:04 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:16:36.599 15:21:04 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@655 -- # es=1 00:16:36.599 15:21:04 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:16:36.599 15:21:04 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:16:36.599 15:21:04 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:16:36.599 15:21:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@779 -- # sleep 1 00:16:37.540 15:21:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:37.540 15:21:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:37.540 15:21:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:37.540 15:21:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:37.540 15:21:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:37.540 15:21:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:37.540 15:21:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:37.540 15:21:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:37.540 15:21:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:37.540 15:21:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:37.540 15:21:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:37.540 15:21:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:37.540 15:21:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:37.540 15:21:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:37.540 15:21:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:37.540 15:21:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:37.540 "name": "raid_bdev1", 00:16:37.540 "uuid": "27b5aa1d-72fc-452e-b440-88de9a7586b4", 00:16:37.540 "strip_size_kb": 0, 00:16:37.540 "state": "online", 00:16:37.540 "raid_level": "raid1", 00:16:37.540 "superblock": true, 00:16:37.540 "num_base_bdevs": 2, 00:16:37.540 "num_base_bdevs_discovered": 1, 00:16:37.540 "num_base_bdevs_operational": 1, 00:16:37.540 "base_bdevs_list": [ 00:16:37.540 { 00:16:37.540 "name": null, 00:16:37.540 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:37.540 "is_configured": false, 00:16:37.540 "data_offset": 0, 00:16:37.540 "data_size": 7936 00:16:37.540 }, 00:16:37.540 { 00:16:37.540 "name": "BaseBdev2", 00:16:37.540 "uuid": "1a122c50-ecfc-5f4e-aec7-6d485817f8d5", 00:16:37.540 "is_configured": true, 00:16:37.540 "data_offset": 256, 00:16:37.540 "data_size": 7936 00:16:37.540 } 00:16:37.540 ] 00:16:37.540 }' 00:16:37.540 15:21:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:37.540 15:21:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:38.111 15:21:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:38.111 15:21:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:38.111 15:21:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:38.111 15:21:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:38.111 15:21:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:38.111 15:21:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:38.112 15:21:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:38.112 15:21:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:38.112 15:21:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:38.112 15:21:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:38.112 15:21:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:38.112 "name": "raid_bdev1", 00:16:38.112 "uuid": "27b5aa1d-72fc-452e-b440-88de9a7586b4", 00:16:38.112 "strip_size_kb": 0, 00:16:38.112 "state": "online", 00:16:38.112 "raid_level": "raid1", 00:16:38.112 "superblock": true, 00:16:38.112 "num_base_bdevs": 2, 00:16:38.112 "num_base_bdevs_discovered": 1, 00:16:38.112 "num_base_bdevs_operational": 1, 00:16:38.112 "base_bdevs_list": [ 00:16:38.112 { 00:16:38.112 "name": null, 00:16:38.112 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:38.112 "is_configured": false, 00:16:38.112 "data_offset": 0, 00:16:38.112 "data_size": 7936 00:16:38.112 }, 00:16:38.112 { 00:16:38.112 "name": "BaseBdev2", 00:16:38.112 "uuid": "1a122c50-ecfc-5f4e-aec7-6d485817f8d5", 00:16:38.112 "is_configured": true, 00:16:38.112 "data_offset": 256, 00:16:38.112 "data_size": 7936 00:16:38.112 } 00:16:38.112 ] 00:16:38.112 }' 00:16:38.112 15:21:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:38.112 15:21:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:38.112 15:21:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:38.112 15:21:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:38.112 15:21:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@784 -- # killprocess 98302 00:16:38.112 15:21:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@954 -- # '[' -z 98302 ']' 00:16:38.112 15:21:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@958 -- # kill -0 98302 00:16:38.112 15:21:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@959 -- # uname 00:16:38.112 15:21:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:38.112 15:21:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 98302 00:16:38.112 15:21:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:38.112 killing process with pid 98302 00:16:38.112 15:21:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:38.112 15:21:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@972 -- # echo 'killing process with pid 98302' 00:16:38.112 15:21:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@973 -- # kill 98302 00:16:38.112 Received shutdown signal, test time was about 60.000000 seconds 00:16:38.112 00:16:38.112 Latency(us) 00:16:38.112 [2024-11-27T15:21:06.219Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:16:38.112 [2024-11-27T15:21:06.219Z] =================================================================================================================== 00:16:38.112 [2024-11-27T15:21:06.219Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:16:38.112 [2024-11-27 15:21:06.136608] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:38.112 15:21:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@978 -- # wait 98302 00:16:38.112 [2024-11-27 15:21:06.136764] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:38.112 [2024-11-27 15:21:06.136813] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:38.112 [2024-11-27 15:21:06.136821] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state offline 00:16:38.112 [2024-11-27 15:21:06.171111] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:38.375 15:21:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@786 -- # return 0 00:16:38.375 00:16:38.375 real 0m18.559s 00:16:38.375 user 0m24.689s 00:16:38.375 sys 0m2.756s 00:16:38.375 15:21:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:38.375 15:21:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:38.375 ************************************ 00:16:38.375 END TEST raid_rebuild_test_sb_md_separate 00:16:38.375 ************************************ 00:16:38.375 15:21:06 bdev_raid -- bdev/bdev_raid.sh@1010 -- # base_malloc_params='-m 32 -i' 00:16:38.375 15:21:06 bdev_raid -- bdev/bdev_raid.sh@1011 -- # run_test raid_state_function_test_sb_md_interleaved raid_state_function_test raid1 2 true 00:16:38.375 15:21:06 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:16:38.375 15:21:06 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:38.375 15:21:06 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:38.375 ************************************ 00:16:38.375 START TEST raid_state_function_test_sb_md_interleaved 00:16:38.375 ************************************ 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 true 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # local strip_size 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@229 -- # raid_pid=98982 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:16:38.375 Process raid pid: 98982 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 98982' 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@231 -- # waitforlisten 98982 00:16:38.375 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@835 -- # '[' -z 98982 ']' 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:38.375 15:21:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:38.653 [2024-11-27 15:21:06.564858] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:16:38.653 [2024-11-27 15:21:06.565085] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:16:38.653 [2024-11-27 15:21:06.741640] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:38.971 [2024-11-27 15:21:06.768948] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:38.971 [2024-11-27 15:21:06.812558] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:38.971 [2024-11-27 15:21:06.812663] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:39.540 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:39.540 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@868 -- # return 0 00:16:39.540 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:39.540 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:39.540 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:39.540 [2024-11-27 15:21:07.383997] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:16:39.540 [2024-11-27 15:21:07.384098] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:16:39.540 [2024-11-27 15:21:07.384128] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:39.540 [2024-11-27 15:21:07.384166] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:39.540 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:39.540 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:39.540 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:39.540 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:39.540 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:39.540 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:39.540 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:39.540 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:39.540 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:39.540 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:39.540 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:39.540 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:39.540 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:39.540 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:39.540 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:39.540 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:39.540 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:39.540 "name": "Existed_Raid", 00:16:39.540 "uuid": "0cad05b0-fbb5-44d2-990a-7081dc5ad656", 00:16:39.540 "strip_size_kb": 0, 00:16:39.540 "state": "configuring", 00:16:39.540 "raid_level": "raid1", 00:16:39.540 "superblock": true, 00:16:39.540 "num_base_bdevs": 2, 00:16:39.540 "num_base_bdevs_discovered": 0, 00:16:39.540 "num_base_bdevs_operational": 2, 00:16:39.540 "base_bdevs_list": [ 00:16:39.540 { 00:16:39.540 "name": "BaseBdev1", 00:16:39.540 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:39.540 "is_configured": false, 00:16:39.540 "data_offset": 0, 00:16:39.540 "data_size": 0 00:16:39.540 }, 00:16:39.540 { 00:16:39.540 "name": "BaseBdev2", 00:16:39.540 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:39.540 "is_configured": false, 00:16:39.540 "data_offset": 0, 00:16:39.540 "data_size": 0 00:16:39.540 } 00:16:39.540 ] 00:16:39.540 }' 00:16:39.540 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:39.540 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:39.800 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:16:39.800 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:39.800 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:39.800 [2024-11-27 15:21:07.847089] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:16:39.800 [2024-11-27 15:21:07.847124] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:16:39.800 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:39.800 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:39.800 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:39.800 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:39.800 [2024-11-27 15:21:07.859095] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:16:39.800 [2024-11-27 15:21:07.859167] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:16:39.800 [2024-11-27 15:21:07.859209] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:39.800 [2024-11-27 15:21:07.859231] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:39.800 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:39.800 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev1 00:16:39.800 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:39.800 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:39.800 [2024-11-27 15:21:07.880017] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:39.800 BaseBdev1 00:16:39.800 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:39.800 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:16:39.800 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:16:39.800 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:16:39.800 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@905 -- # local i 00:16:39.800 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:16:39.800 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:16:39.800 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:16:39.800 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:39.800 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:39.800 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:39.800 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:16:39.800 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:39.800 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:40.061 [ 00:16:40.061 { 00:16:40.061 "name": "BaseBdev1", 00:16:40.061 "aliases": [ 00:16:40.061 "0f790105-63ec-4704-83f7-b6614def8811" 00:16:40.061 ], 00:16:40.061 "product_name": "Malloc disk", 00:16:40.061 "block_size": 4128, 00:16:40.061 "num_blocks": 8192, 00:16:40.061 "uuid": "0f790105-63ec-4704-83f7-b6614def8811", 00:16:40.061 "md_size": 32, 00:16:40.061 "md_interleave": true, 00:16:40.061 "dif_type": 0, 00:16:40.061 "assigned_rate_limits": { 00:16:40.061 "rw_ios_per_sec": 0, 00:16:40.061 "rw_mbytes_per_sec": 0, 00:16:40.061 "r_mbytes_per_sec": 0, 00:16:40.061 "w_mbytes_per_sec": 0 00:16:40.061 }, 00:16:40.061 "claimed": true, 00:16:40.061 "claim_type": "exclusive_write", 00:16:40.061 "zoned": false, 00:16:40.061 "supported_io_types": { 00:16:40.061 "read": true, 00:16:40.061 "write": true, 00:16:40.061 "unmap": true, 00:16:40.061 "flush": true, 00:16:40.061 "reset": true, 00:16:40.061 "nvme_admin": false, 00:16:40.061 "nvme_io": false, 00:16:40.061 "nvme_io_md": false, 00:16:40.061 "write_zeroes": true, 00:16:40.061 "zcopy": true, 00:16:40.061 "get_zone_info": false, 00:16:40.061 "zone_management": false, 00:16:40.061 "zone_append": false, 00:16:40.061 "compare": false, 00:16:40.061 "compare_and_write": false, 00:16:40.061 "abort": true, 00:16:40.061 "seek_hole": false, 00:16:40.061 "seek_data": false, 00:16:40.061 "copy": true, 00:16:40.061 "nvme_iov_md": false 00:16:40.061 }, 00:16:40.061 "memory_domains": [ 00:16:40.061 { 00:16:40.061 "dma_device_id": "system", 00:16:40.061 "dma_device_type": 1 00:16:40.061 }, 00:16:40.061 { 00:16:40.061 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:40.061 "dma_device_type": 2 00:16:40.061 } 00:16:40.061 ], 00:16:40.061 "driver_specific": {} 00:16:40.061 } 00:16:40.061 ] 00:16:40.061 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:40.061 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@911 -- # return 0 00:16:40.061 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:40.061 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:40.061 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:40.061 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:40.061 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:40.061 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:40.061 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:40.061 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:40.061 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:40.061 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:40.061 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:40.061 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:40.061 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:40.061 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:40.061 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:40.061 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:40.061 "name": "Existed_Raid", 00:16:40.061 "uuid": "91844f4e-118c-4a35-a65d-cdc87d6da2dd", 00:16:40.061 "strip_size_kb": 0, 00:16:40.061 "state": "configuring", 00:16:40.061 "raid_level": "raid1", 00:16:40.061 "superblock": true, 00:16:40.061 "num_base_bdevs": 2, 00:16:40.061 "num_base_bdevs_discovered": 1, 00:16:40.061 "num_base_bdevs_operational": 2, 00:16:40.061 "base_bdevs_list": [ 00:16:40.061 { 00:16:40.061 "name": "BaseBdev1", 00:16:40.061 "uuid": "0f790105-63ec-4704-83f7-b6614def8811", 00:16:40.061 "is_configured": true, 00:16:40.061 "data_offset": 256, 00:16:40.061 "data_size": 7936 00:16:40.061 }, 00:16:40.061 { 00:16:40.061 "name": "BaseBdev2", 00:16:40.061 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:40.061 "is_configured": false, 00:16:40.061 "data_offset": 0, 00:16:40.061 "data_size": 0 00:16:40.061 } 00:16:40.061 ] 00:16:40.061 }' 00:16:40.061 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:40.061 15:21:07 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:40.321 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:16:40.321 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:40.321 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:40.321 [2024-11-27 15:21:08.343589] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:16:40.321 [2024-11-27 15:21:08.343678] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:16:40.321 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:40.321 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:40.321 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:40.321 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:40.321 [2024-11-27 15:21:08.355575] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:40.321 [2024-11-27 15:21:08.357420] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:40.321 [2024-11-27 15:21:08.357498] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:40.321 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:40.321 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:16:40.321 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:16:40.321 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:40.321 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:40.321 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:40.321 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:40.321 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:40.321 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:40.321 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:40.322 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:40.322 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:40.322 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:40.322 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:40.322 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:40.322 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:40.322 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:40.322 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:40.322 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:40.322 "name": "Existed_Raid", 00:16:40.322 "uuid": "1f27b6c5-1a42-42c7-bc0d-5784db8964da", 00:16:40.322 "strip_size_kb": 0, 00:16:40.322 "state": "configuring", 00:16:40.322 "raid_level": "raid1", 00:16:40.322 "superblock": true, 00:16:40.322 "num_base_bdevs": 2, 00:16:40.322 "num_base_bdevs_discovered": 1, 00:16:40.322 "num_base_bdevs_operational": 2, 00:16:40.322 "base_bdevs_list": [ 00:16:40.322 { 00:16:40.322 "name": "BaseBdev1", 00:16:40.322 "uuid": "0f790105-63ec-4704-83f7-b6614def8811", 00:16:40.322 "is_configured": true, 00:16:40.322 "data_offset": 256, 00:16:40.322 "data_size": 7936 00:16:40.322 }, 00:16:40.322 { 00:16:40.322 "name": "BaseBdev2", 00:16:40.322 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:40.322 "is_configured": false, 00:16:40.322 "data_offset": 0, 00:16:40.322 "data_size": 0 00:16:40.322 } 00:16:40.322 ] 00:16:40.322 }' 00:16:40.322 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:40.322 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev2 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:40.893 [2024-11-27 15:21:08.762168] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:40.893 [2024-11-27 15:21:08.762421] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:16:40.893 [2024-11-27 15:21:08.762478] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:40.893 [2024-11-27 15:21:08.762593] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:16:40.893 [2024-11-27 15:21:08.762703] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:16:40.893 [2024-11-27 15:21:08.762747] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:16:40.893 BaseBdev2 00:16:40.893 [2024-11-27 15:21:08.762837] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@905 -- # local i 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:40.893 [ 00:16:40.893 { 00:16:40.893 "name": "BaseBdev2", 00:16:40.893 "aliases": [ 00:16:40.893 "84446e02-258a-401c-8621-458409c5a868" 00:16:40.893 ], 00:16:40.893 "product_name": "Malloc disk", 00:16:40.893 "block_size": 4128, 00:16:40.893 "num_blocks": 8192, 00:16:40.893 "uuid": "84446e02-258a-401c-8621-458409c5a868", 00:16:40.893 "md_size": 32, 00:16:40.893 "md_interleave": true, 00:16:40.893 "dif_type": 0, 00:16:40.893 "assigned_rate_limits": { 00:16:40.893 "rw_ios_per_sec": 0, 00:16:40.893 "rw_mbytes_per_sec": 0, 00:16:40.893 "r_mbytes_per_sec": 0, 00:16:40.893 "w_mbytes_per_sec": 0 00:16:40.893 }, 00:16:40.893 "claimed": true, 00:16:40.893 "claim_type": "exclusive_write", 00:16:40.893 "zoned": false, 00:16:40.893 "supported_io_types": { 00:16:40.893 "read": true, 00:16:40.893 "write": true, 00:16:40.893 "unmap": true, 00:16:40.893 "flush": true, 00:16:40.893 "reset": true, 00:16:40.893 "nvme_admin": false, 00:16:40.893 "nvme_io": false, 00:16:40.893 "nvme_io_md": false, 00:16:40.893 "write_zeroes": true, 00:16:40.893 "zcopy": true, 00:16:40.893 "get_zone_info": false, 00:16:40.893 "zone_management": false, 00:16:40.893 "zone_append": false, 00:16:40.893 "compare": false, 00:16:40.893 "compare_and_write": false, 00:16:40.893 "abort": true, 00:16:40.893 "seek_hole": false, 00:16:40.893 "seek_data": false, 00:16:40.893 "copy": true, 00:16:40.893 "nvme_iov_md": false 00:16:40.893 }, 00:16:40.893 "memory_domains": [ 00:16:40.893 { 00:16:40.893 "dma_device_id": "system", 00:16:40.893 "dma_device_type": 1 00:16:40.893 }, 00:16:40.893 { 00:16:40.893 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:40.893 "dma_device_type": 2 00:16:40.893 } 00:16:40.893 ], 00:16:40.893 "driver_specific": {} 00:16:40.893 } 00:16:40.893 ] 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@911 -- # return 0 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:40.893 "name": "Existed_Raid", 00:16:40.893 "uuid": "1f27b6c5-1a42-42c7-bc0d-5784db8964da", 00:16:40.893 "strip_size_kb": 0, 00:16:40.893 "state": "online", 00:16:40.893 "raid_level": "raid1", 00:16:40.893 "superblock": true, 00:16:40.893 "num_base_bdevs": 2, 00:16:40.893 "num_base_bdevs_discovered": 2, 00:16:40.893 "num_base_bdevs_operational": 2, 00:16:40.893 "base_bdevs_list": [ 00:16:40.893 { 00:16:40.893 "name": "BaseBdev1", 00:16:40.893 "uuid": "0f790105-63ec-4704-83f7-b6614def8811", 00:16:40.893 "is_configured": true, 00:16:40.893 "data_offset": 256, 00:16:40.893 "data_size": 7936 00:16:40.893 }, 00:16:40.893 { 00:16:40.893 "name": "BaseBdev2", 00:16:40.893 "uuid": "84446e02-258a-401c-8621-458409c5a868", 00:16:40.893 "is_configured": true, 00:16:40.893 "data_offset": 256, 00:16:40.893 "data_size": 7936 00:16:40.893 } 00:16:40.893 ] 00:16:40.893 }' 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:40.893 15:21:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:41.154 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:16:41.154 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:16:41.154 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:41.154 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:41.154 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:16:41.154 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:41.154 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:16:41.154 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:41.154 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:41.154 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:41.154 [2024-11-27 15:21:09.229693] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:41.154 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:41.154 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:41.154 "name": "Existed_Raid", 00:16:41.154 "aliases": [ 00:16:41.154 "1f27b6c5-1a42-42c7-bc0d-5784db8964da" 00:16:41.154 ], 00:16:41.154 "product_name": "Raid Volume", 00:16:41.154 "block_size": 4128, 00:16:41.154 "num_blocks": 7936, 00:16:41.154 "uuid": "1f27b6c5-1a42-42c7-bc0d-5784db8964da", 00:16:41.154 "md_size": 32, 00:16:41.154 "md_interleave": true, 00:16:41.154 "dif_type": 0, 00:16:41.154 "assigned_rate_limits": { 00:16:41.154 "rw_ios_per_sec": 0, 00:16:41.154 "rw_mbytes_per_sec": 0, 00:16:41.154 "r_mbytes_per_sec": 0, 00:16:41.154 "w_mbytes_per_sec": 0 00:16:41.154 }, 00:16:41.154 "claimed": false, 00:16:41.154 "zoned": false, 00:16:41.154 "supported_io_types": { 00:16:41.154 "read": true, 00:16:41.154 "write": true, 00:16:41.154 "unmap": false, 00:16:41.154 "flush": false, 00:16:41.154 "reset": true, 00:16:41.154 "nvme_admin": false, 00:16:41.154 "nvme_io": false, 00:16:41.154 "nvme_io_md": false, 00:16:41.154 "write_zeroes": true, 00:16:41.154 "zcopy": false, 00:16:41.154 "get_zone_info": false, 00:16:41.154 "zone_management": false, 00:16:41.154 "zone_append": false, 00:16:41.154 "compare": false, 00:16:41.154 "compare_and_write": false, 00:16:41.154 "abort": false, 00:16:41.154 "seek_hole": false, 00:16:41.154 "seek_data": false, 00:16:41.154 "copy": false, 00:16:41.154 "nvme_iov_md": false 00:16:41.154 }, 00:16:41.154 "memory_domains": [ 00:16:41.154 { 00:16:41.154 "dma_device_id": "system", 00:16:41.154 "dma_device_type": 1 00:16:41.154 }, 00:16:41.154 { 00:16:41.154 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:41.154 "dma_device_type": 2 00:16:41.154 }, 00:16:41.154 { 00:16:41.154 "dma_device_id": "system", 00:16:41.154 "dma_device_type": 1 00:16:41.154 }, 00:16:41.154 { 00:16:41.154 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:41.154 "dma_device_type": 2 00:16:41.154 } 00:16:41.154 ], 00:16:41.154 "driver_specific": { 00:16:41.154 "raid": { 00:16:41.154 "uuid": "1f27b6c5-1a42-42c7-bc0d-5784db8964da", 00:16:41.154 "strip_size_kb": 0, 00:16:41.154 "state": "online", 00:16:41.154 "raid_level": "raid1", 00:16:41.154 "superblock": true, 00:16:41.154 "num_base_bdevs": 2, 00:16:41.154 "num_base_bdevs_discovered": 2, 00:16:41.154 "num_base_bdevs_operational": 2, 00:16:41.154 "base_bdevs_list": [ 00:16:41.154 { 00:16:41.154 "name": "BaseBdev1", 00:16:41.154 "uuid": "0f790105-63ec-4704-83f7-b6614def8811", 00:16:41.154 "is_configured": true, 00:16:41.154 "data_offset": 256, 00:16:41.154 "data_size": 7936 00:16:41.154 }, 00:16:41.154 { 00:16:41.154 "name": "BaseBdev2", 00:16:41.155 "uuid": "84446e02-258a-401c-8621-458409c5a868", 00:16:41.155 "is_configured": true, 00:16:41.155 "data_offset": 256, 00:16:41.155 "data_size": 7936 00:16:41.155 } 00:16:41.155 ] 00:16:41.155 } 00:16:41.155 } 00:16:41.155 }' 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:16:41.415 BaseBdev2' 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:41.415 [2024-11-27 15:21:09.445094] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@260 -- # local expected_state 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@198 -- # case $1 in 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@199 -- # return 0 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:41.415 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:41.416 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:41.416 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:41.416 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:41.416 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:41.416 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:41.416 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:41.416 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:41.416 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:41.416 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:41.416 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:41.416 "name": "Existed_Raid", 00:16:41.416 "uuid": "1f27b6c5-1a42-42c7-bc0d-5784db8964da", 00:16:41.416 "strip_size_kb": 0, 00:16:41.416 "state": "online", 00:16:41.416 "raid_level": "raid1", 00:16:41.416 "superblock": true, 00:16:41.416 "num_base_bdevs": 2, 00:16:41.416 "num_base_bdevs_discovered": 1, 00:16:41.416 "num_base_bdevs_operational": 1, 00:16:41.416 "base_bdevs_list": [ 00:16:41.416 { 00:16:41.416 "name": null, 00:16:41.416 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:41.416 "is_configured": false, 00:16:41.416 "data_offset": 0, 00:16:41.416 "data_size": 7936 00:16:41.416 }, 00:16:41.416 { 00:16:41.416 "name": "BaseBdev2", 00:16:41.416 "uuid": "84446e02-258a-401c-8621-458409c5a868", 00:16:41.416 "is_configured": true, 00:16:41.416 "data_offset": 256, 00:16:41.416 "data_size": 7936 00:16:41.416 } 00:16:41.416 ] 00:16:41.416 }' 00:16:41.416 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:41.416 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:41.987 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:16:41.987 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:16:41.987 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:16:41.987 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:41.987 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:41.987 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:41.987 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:41.987 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:16:41.987 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:16:41.987 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:16:41.987 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:41.987 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:41.987 [2024-11-27 15:21:09.932051] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:16:41.987 [2024-11-27 15:21:09.932190] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:41.987 [2024-11-27 15:21:09.944319] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:41.987 [2024-11-27 15:21:09.944456] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:41.987 [2024-11-27 15:21:09.944496] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:16:41.987 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:41.987 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:16:41.987 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:16:41.987 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:41.987 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:16:41.987 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:41.987 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:41.987 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:41.987 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:16:41.987 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:16:41.987 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:16:41.987 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@326 -- # killprocess 98982 00:16:41.987 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@954 -- # '[' -z 98982 ']' 00:16:41.987 15:21:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@958 -- # kill -0 98982 00:16:41.987 15:21:10 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@959 -- # uname 00:16:41.987 15:21:10 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:41.987 15:21:10 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 98982 00:16:41.987 15:21:10 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:41.987 15:21:10 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:41.987 killing process with pid 98982 00:16:41.987 15:21:10 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@972 -- # echo 'killing process with pid 98982' 00:16:41.987 15:21:10 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@973 -- # kill 98982 00:16:41.987 [2024-11-27 15:21:10.042426] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:41.987 15:21:10 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@978 -- # wait 98982 00:16:41.988 [2024-11-27 15:21:10.043389] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:42.248 15:21:10 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@328 -- # return 0 00:16:42.248 00:16:42.248 real 0m3.807s 00:16:42.248 user 0m5.921s 00:16:42.248 sys 0m0.893s 00:16:42.248 15:21:10 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:42.248 ************************************ 00:16:42.248 END TEST raid_state_function_test_sb_md_interleaved 00:16:42.248 ************************************ 00:16:42.248 15:21:10 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:42.248 15:21:10 bdev_raid -- bdev/bdev_raid.sh@1012 -- # run_test raid_superblock_test_md_interleaved raid_superblock_test raid1 2 00:16:42.248 15:21:10 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:16:42.248 15:21:10 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:42.248 15:21:10 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:42.248 ************************************ 00:16:42.248 START TEST raid_superblock_test_md_interleaved 00:16:42.248 ************************************ 00:16:42.248 15:21:10 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 2 00:16:42.248 15:21:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:16:42.248 15:21:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:16:42.248 15:21:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:16:42.248 15:21:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:16:42.248 15:21:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:16:42.248 15:21:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:16:42.248 15:21:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:16:42.248 15:21:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:16:42.248 15:21:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:16:42.248 15:21:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@399 -- # local strip_size 00:16:42.248 15:21:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:16:42.248 15:21:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:16:42.248 15:21:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:16:42.248 15:21:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:16:42.248 15:21:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:16:42.248 15:21:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@412 -- # raid_pid=99223 00:16:42.248 15:21:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:16:42.248 15:21:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@413 -- # waitforlisten 99223 00:16:42.248 15:21:10 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@835 -- # '[' -z 99223 ']' 00:16:42.248 15:21:10 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:42.248 15:21:10 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:42.248 15:21:10 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:42.248 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:42.248 15:21:10 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:42.248 15:21:10 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:42.508 [2024-11-27 15:21:10.438254] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:16:42.508 [2024-11-27 15:21:10.438405] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid99223 ] 00:16:42.768 [2024-11-27 15:21:10.614614] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:42.768 [2024-11-27 15:21:10.641325] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:42.768 [2024-11-27 15:21:10.684144] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:42.768 [2024-11-27 15:21:10.684182] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@868 -- # return 0 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b malloc1 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:43.340 malloc1 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:43.340 [2024-11-27 15:21:11.283756] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:43.340 [2024-11-27 15:21:11.283870] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:43.340 [2024-11-27 15:21:11.283924] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:16:43.340 [2024-11-27 15:21:11.283983] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:43.340 [2024-11-27 15:21:11.285802] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:43.340 [2024-11-27 15:21:11.285876] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:43.340 pt1 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b malloc2 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:43.340 malloc2 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:43.340 [2024-11-27 15:21:11.312759] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:43.340 [2024-11-27 15:21:11.312840] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:43.340 [2024-11-27 15:21:11.312856] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:16:43.340 [2024-11-27 15:21:11.312866] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:43.340 [2024-11-27 15:21:11.314884] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:43.340 [2024-11-27 15:21:11.314930] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:43.340 pt2 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:43.340 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:16:43.341 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:43.341 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:16:43.341 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:43.341 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:43.341 [2024-11-27 15:21:11.324775] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:43.341 [2024-11-27 15:21:11.326622] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:43.341 [2024-11-27 15:21:11.326858] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:16:43.341 [2024-11-27 15:21:11.326877] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:43.341 [2024-11-27 15:21:11.326985] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:16:43.341 [2024-11-27 15:21:11.327063] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:16:43.341 [2024-11-27 15:21:11.327074] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:16:43.341 [2024-11-27 15:21:11.327141] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:43.341 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:43.341 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:43.341 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:43.341 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:43.341 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:43.341 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:43.341 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:43.341 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:43.341 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:43.341 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:43.341 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:43.341 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:43.341 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:43.341 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:43.341 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:43.341 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:43.341 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:43.341 "name": "raid_bdev1", 00:16:43.341 "uuid": "cc1e72b4-aa2b-41e9-b157-0ba7845dc144", 00:16:43.341 "strip_size_kb": 0, 00:16:43.341 "state": "online", 00:16:43.341 "raid_level": "raid1", 00:16:43.341 "superblock": true, 00:16:43.341 "num_base_bdevs": 2, 00:16:43.341 "num_base_bdevs_discovered": 2, 00:16:43.341 "num_base_bdevs_operational": 2, 00:16:43.341 "base_bdevs_list": [ 00:16:43.341 { 00:16:43.341 "name": "pt1", 00:16:43.341 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:43.341 "is_configured": true, 00:16:43.341 "data_offset": 256, 00:16:43.341 "data_size": 7936 00:16:43.341 }, 00:16:43.341 { 00:16:43.341 "name": "pt2", 00:16:43.341 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:43.341 "is_configured": true, 00:16:43.341 "data_offset": 256, 00:16:43.341 "data_size": 7936 00:16:43.341 } 00:16:43.341 ] 00:16:43.341 }' 00:16:43.341 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:43.341 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:43.911 [2024-11-27 15:21:11.772286] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:43.911 "name": "raid_bdev1", 00:16:43.911 "aliases": [ 00:16:43.911 "cc1e72b4-aa2b-41e9-b157-0ba7845dc144" 00:16:43.911 ], 00:16:43.911 "product_name": "Raid Volume", 00:16:43.911 "block_size": 4128, 00:16:43.911 "num_blocks": 7936, 00:16:43.911 "uuid": "cc1e72b4-aa2b-41e9-b157-0ba7845dc144", 00:16:43.911 "md_size": 32, 00:16:43.911 "md_interleave": true, 00:16:43.911 "dif_type": 0, 00:16:43.911 "assigned_rate_limits": { 00:16:43.911 "rw_ios_per_sec": 0, 00:16:43.911 "rw_mbytes_per_sec": 0, 00:16:43.911 "r_mbytes_per_sec": 0, 00:16:43.911 "w_mbytes_per_sec": 0 00:16:43.911 }, 00:16:43.911 "claimed": false, 00:16:43.911 "zoned": false, 00:16:43.911 "supported_io_types": { 00:16:43.911 "read": true, 00:16:43.911 "write": true, 00:16:43.911 "unmap": false, 00:16:43.911 "flush": false, 00:16:43.911 "reset": true, 00:16:43.911 "nvme_admin": false, 00:16:43.911 "nvme_io": false, 00:16:43.911 "nvme_io_md": false, 00:16:43.911 "write_zeroes": true, 00:16:43.911 "zcopy": false, 00:16:43.911 "get_zone_info": false, 00:16:43.911 "zone_management": false, 00:16:43.911 "zone_append": false, 00:16:43.911 "compare": false, 00:16:43.911 "compare_and_write": false, 00:16:43.911 "abort": false, 00:16:43.911 "seek_hole": false, 00:16:43.911 "seek_data": false, 00:16:43.911 "copy": false, 00:16:43.911 "nvme_iov_md": false 00:16:43.911 }, 00:16:43.911 "memory_domains": [ 00:16:43.911 { 00:16:43.911 "dma_device_id": "system", 00:16:43.911 "dma_device_type": 1 00:16:43.911 }, 00:16:43.911 { 00:16:43.911 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:43.911 "dma_device_type": 2 00:16:43.911 }, 00:16:43.911 { 00:16:43.911 "dma_device_id": "system", 00:16:43.911 "dma_device_type": 1 00:16:43.911 }, 00:16:43.911 { 00:16:43.911 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:43.911 "dma_device_type": 2 00:16:43.911 } 00:16:43.911 ], 00:16:43.911 "driver_specific": { 00:16:43.911 "raid": { 00:16:43.911 "uuid": "cc1e72b4-aa2b-41e9-b157-0ba7845dc144", 00:16:43.911 "strip_size_kb": 0, 00:16:43.911 "state": "online", 00:16:43.911 "raid_level": "raid1", 00:16:43.911 "superblock": true, 00:16:43.911 "num_base_bdevs": 2, 00:16:43.911 "num_base_bdevs_discovered": 2, 00:16:43.911 "num_base_bdevs_operational": 2, 00:16:43.911 "base_bdevs_list": [ 00:16:43.911 { 00:16:43.911 "name": "pt1", 00:16:43.911 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:43.911 "is_configured": true, 00:16:43.911 "data_offset": 256, 00:16:43.911 "data_size": 7936 00:16:43.911 }, 00:16:43.911 { 00:16:43.911 "name": "pt2", 00:16:43.911 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:43.911 "is_configured": true, 00:16:43.911 "data_offset": 256, 00:16:43.911 "data_size": 7936 00:16:43.911 } 00:16:43.911 ] 00:16:43.911 } 00:16:43.911 } 00:16:43.911 }' 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:16:43.911 pt2' 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:16:43.911 [2024-11-27 15:21:11.987824] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:43.911 15:21:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=cc1e72b4-aa2b-41e9-b157-0ba7845dc144 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@436 -- # '[' -z cc1e72b4-aa2b-41e9-b157-0ba7845dc144 ']' 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:44.172 [2024-11-27 15:21:12.035555] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:44.172 [2024-11-27 15:21:12.035621] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:44.172 [2024-11-27 15:21:12.035731] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:44.172 [2024-11-27 15:21:12.035812] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:44.172 [2024-11-27 15:21:12.035857] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@652 -- # local es=0 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:44.172 [2024-11-27 15:21:12.175373] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:16:44.172 [2024-11-27 15:21:12.177345] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:16:44.172 [2024-11-27 15:21:12.177462] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:16:44.172 [2024-11-27 15:21:12.177537] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:16:44.172 [2024-11-27 15:21:12.177593] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:44.172 [2024-11-27 15:21:12.177613] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state configuring 00:16:44.172 request: 00:16:44.172 { 00:16:44.172 "name": "raid_bdev1", 00:16:44.172 "raid_level": "raid1", 00:16:44.172 "base_bdevs": [ 00:16:44.172 "malloc1", 00:16:44.172 "malloc2" 00:16:44.172 ], 00:16:44.172 "superblock": false, 00:16:44.172 "method": "bdev_raid_create", 00:16:44.172 "req_id": 1 00:16:44.172 } 00:16:44.172 Got JSON-RPC error response 00:16:44.172 response: 00:16:44.172 { 00:16:44.172 "code": -17, 00:16:44.172 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:16:44.172 } 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@655 -- # es=1 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:44.172 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:44.173 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:44.173 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:44.173 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:16:44.173 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:16:44.173 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:44.173 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:44.173 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:44.173 [2024-11-27 15:21:12.223255] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:44.173 [2024-11-27 15:21:12.223298] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:44.173 [2024-11-27 15:21:12.223330] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:16:44.173 [2024-11-27 15:21:12.223339] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:44.173 [2024-11-27 15:21:12.225242] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:44.173 [2024-11-27 15:21:12.225316] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:44.173 [2024-11-27 15:21:12.225363] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:16:44.173 [2024-11-27 15:21:12.225400] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:44.173 pt1 00:16:44.173 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:44.173 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:16:44.173 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:44.173 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:44.173 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:44.173 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:44.173 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:44.173 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:44.173 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:44.173 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:44.173 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:44.173 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:44.173 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:44.173 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:44.173 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:44.173 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:44.433 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:44.433 "name": "raid_bdev1", 00:16:44.433 "uuid": "cc1e72b4-aa2b-41e9-b157-0ba7845dc144", 00:16:44.433 "strip_size_kb": 0, 00:16:44.433 "state": "configuring", 00:16:44.433 "raid_level": "raid1", 00:16:44.433 "superblock": true, 00:16:44.433 "num_base_bdevs": 2, 00:16:44.433 "num_base_bdevs_discovered": 1, 00:16:44.433 "num_base_bdevs_operational": 2, 00:16:44.433 "base_bdevs_list": [ 00:16:44.433 { 00:16:44.433 "name": "pt1", 00:16:44.433 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:44.433 "is_configured": true, 00:16:44.433 "data_offset": 256, 00:16:44.433 "data_size": 7936 00:16:44.433 }, 00:16:44.433 { 00:16:44.433 "name": null, 00:16:44.433 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:44.433 "is_configured": false, 00:16:44.433 "data_offset": 256, 00:16:44.433 "data_size": 7936 00:16:44.433 } 00:16:44.433 ] 00:16:44.433 }' 00:16:44.433 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:44.433 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:44.693 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:16:44.693 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:16:44.694 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:16:44.694 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:44.694 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:44.694 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:44.694 [2024-11-27 15:21:12.702453] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:44.694 [2024-11-27 15:21:12.702569] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:44.694 [2024-11-27 15:21:12.702610] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:16:44.694 [2024-11-27 15:21:12.702637] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:44.694 [2024-11-27 15:21:12.702789] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:44.694 [2024-11-27 15:21:12.702828] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:44.694 [2024-11-27 15:21:12.702891] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:16:44.694 [2024-11-27 15:21:12.702943] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:44.694 [2024-11-27 15:21:12.703054] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:16:44.694 [2024-11-27 15:21:12.703094] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:44.694 [2024-11-27 15:21:12.703189] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:16:44.694 [2024-11-27 15:21:12.703274] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:16:44.694 [2024-11-27 15:21:12.703311] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:16:44.694 [2024-11-27 15:21:12.703406] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:44.694 pt2 00:16:44.694 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:44.694 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:16:44.694 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:16:44.694 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:44.694 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:44.694 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:44.694 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:44.694 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:44.694 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:44.694 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:44.694 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:44.694 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:44.694 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:44.694 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:44.694 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:44.694 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:44.694 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:44.694 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:44.694 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:44.694 "name": "raid_bdev1", 00:16:44.694 "uuid": "cc1e72b4-aa2b-41e9-b157-0ba7845dc144", 00:16:44.694 "strip_size_kb": 0, 00:16:44.694 "state": "online", 00:16:44.694 "raid_level": "raid1", 00:16:44.694 "superblock": true, 00:16:44.694 "num_base_bdevs": 2, 00:16:44.694 "num_base_bdevs_discovered": 2, 00:16:44.694 "num_base_bdevs_operational": 2, 00:16:44.694 "base_bdevs_list": [ 00:16:44.694 { 00:16:44.694 "name": "pt1", 00:16:44.694 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:44.694 "is_configured": true, 00:16:44.694 "data_offset": 256, 00:16:44.694 "data_size": 7936 00:16:44.694 }, 00:16:44.694 { 00:16:44.694 "name": "pt2", 00:16:44.694 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:44.694 "is_configured": true, 00:16:44.694 "data_offset": 256, 00:16:44.694 "data_size": 7936 00:16:44.694 } 00:16:44.694 ] 00:16:44.694 }' 00:16:44.694 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:44.694 15:21:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:45.262 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:16:45.262 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:16:45.262 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:45.262 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:45.262 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:16:45.262 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:45.262 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:45.262 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:45.262 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:45.262 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:45.262 [2024-11-27 15:21:13.151155] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:45.262 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:45.262 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:45.262 "name": "raid_bdev1", 00:16:45.262 "aliases": [ 00:16:45.262 "cc1e72b4-aa2b-41e9-b157-0ba7845dc144" 00:16:45.262 ], 00:16:45.262 "product_name": "Raid Volume", 00:16:45.262 "block_size": 4128, 00:16:45.262 "num_blocks": 7936, 00:16:45.262 "uuid": "cc1e72b4-aa2b-41e9-b157-0ba7845dc144", 00:16:45.262 "md_size": 32, 00:16:45.262 "md_interleave": true, 00:16:45.262 "dif_type": 0, 00:16:45.262 "assigned_rate_limits": { 00:16:45.262 "rw_ios_per_sec": 0, 00:16:45.262 "rw_mbytes_per_sec": 0, 00:16:45.262 "r_mbytes_per_sec": 0, 00:16:45.262 "w_mbytes_per_sec": 0 00:16:45.262 }, 00:16:45.262 "claimed": false, 00:16:45.262 "zoned": false, 00:16:45.262 "supported_io_types": { 00:16:45.262 "read": true, 00:16:45.262 "write": true, 00:16:45.262 "unmap": false, 00:16:45.262 "flush": false, 00:16:45.262 "reset": true, 00:16:45.262 "nvme_admin": false, 00:16:45.262 "nvme_io": false, 00:16:45.262 "nvme_io_md": false, 00:16:45.262 "write_zeroes": true, 00:16:45.262 "zcopy": false, 00:16:45.262 "get_zone_info": false, 00:16:45.262 "zone_management": false, 00:16:45.262 "zone_append": false, 00:16:45.262 "compare": false, 00:16:45.262 "compare_and_write": false, 00:16:45.262 "abort": false, 00:16:45.262 "seek_hole": false, 00:16:45.262 "seek_data": false, 00:16:45.262 "copy": false, 00:16:45.262 "nvme_iov_md": false 00:16:45.262 }, 00:16:45.262 "memory_domains": [ 00:16:45.262 { 00:16:45.262 "dma_device_id": "system", 00:16:45.262 "dma_device_type": 1 00:16:45.262 }, 00:16:45.262 { 00:16:45.262 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:45.262 "dma_device_type": 2 00:16:45.262 }, 00:16:45.262 { 00:16:45.262 "dma_device_id": "system", 00:16:45.262 "dma_device_type": 1 00:16:45.262 }, 00:16:45.262 { 00:16:45.262 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:45.262 "dma_device_type": 2 00:16:45.262 } 00:16:45.262 ], 00:16:45.262 "driver_specific": { 00:16:45.262 "raid": { 00:16:45.262 "uuid": "cc1e72b4-aa2b-41e9-b157-0ba7845dc144", 00:16:45.262 "strip_size_kb": 0, 00:16:45.262 "state": "online", 00:16:45.262 "raid_level": "raid1", 00:16:45.262 "superblock": true, 00:16:45.262 "num_base_bdevs": 2, 00:16:45.262 "num_base_bdevs_discovered": 2, 00:16:45.262 "num_base_bdevs_operational": 2, 00:16:45.262 "base_bdevs_list": [ 00:16:45.262 { 00:16:45.262 "name": "pt1", 00:16:45.262 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:45.262 "is_configured": true, 00:16:45.262 "data_offset": 256, 00:16:45.262 "data_size": 7936 00:16:45.262 }, 00:16:45.262 { 00:16:45.262 "name": "pt2", 00:16:45.262 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:45.262 "is_configured": true, 00:16:45.262 "data_offset": 256, 00:16:45.262 "data_size": 7936 00:16:45.262 } 00:16:45.262 ] 00:16:45.262 } 00:16:45.262 } 00:16:45.262 }' 00:16:45.262 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:45.262 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:16:45.262 pt2' 00:16:45.262 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:45.262 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:16:45.262 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:45.262 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:45.262 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:16:45.262 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:45.262 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:45.262 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:45.262 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:45.262 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:45.262 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:45.262 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:45.262 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:16:45.262 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:45.262 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:45.262 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:45.521 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:45.522 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:45.522 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:45.522 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:16:45.522 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:45.522 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:45.522 [2024-11-27 15:21:13.394533] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:45.522 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:45.522 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # '[' cc1e72b4-aa2b-41e9-b157-0ba7845dc144 '!=' cc1e72b4-aa2b-41e9-b157-0ba7845dc144 ']' 00:16:45.522 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:16:45.522 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@198 -- # case $1 in 00:16:45.522 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@199 -- # return 0 00:16:45.522 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:16:45.522 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:45.522 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:45.522 [2024-11-27 15:21:13.442253] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:16:45.522 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:45.522 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:45.522 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:45.522 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:45.522 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:45.522 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:45.522 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:45.522 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:45.522 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:45.522 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:45.522 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:45.522 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:45.522 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:45.522 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:45.522 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:45.522 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:45.522 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:45.522 "name": "raid_bdev1", 00:16:45.522 "uuid": "cc1e72b4-aa2b-41e9-b157-0ba7845dc144", 00:16:45.522 "strip_size_kb": 0, 00:16:45.522 "state": "online", 00:16:45.522 "raid_level": "raid1", 00:16:45.522 "superblock": true, 00:16:45.522 "num_base_bdevs": 2, 00:16:45.522 "num_base_bdevs_discovered": 1, 00:16:45.522 "num_base_bdevs_operational": 1, 00:16:45.522 "base_bdevs_list": [ 00:16:45.522 { 00:16:45.522 "name": null, 00:16:45.522 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:45.522 "is_configured": false, 00:16:45.522 "data_offset": 0, 00:16:45.522 "data_size": 7936 00:16:45.522 }, 00:16:45.522 { 00:16:45.522 "name": "pt2", 00:16:45.522 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:45.522 "is_configured": true, 00:16:45.522 "data_offset": 256, 00:16:45.522 "data_size": 7936 00:16:45.522 } 00:16:45.522 ] 00:16:45.522 }' 00:16:45.522 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:45.522 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:46.091 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:46.092 [2024-11-27 15:21:13.893374] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:46.092 [2024-11-27 15:21:13.893454] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:46.092 [2024-11-27 15:21:13.893559] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:46.092 [2024-11-27 15:21:13.893645] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:46.092 [2024-11-27 15:21:13.893696] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@519 -- # i=1 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:46.092 [2024-11-27 15:21:13.949322] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:46.092 [2024-11-27 15:21:13.949376] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:46.092 [2024-11-27 15:21:13.949399] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:16:46.092 [2024-11-27 15:21:13.949409] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:46.092 [2024-11-27 15:21:13.951557] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:46.092 [2024-11-27 15:21:13.951598] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:46.092 [2024-11-27 15:21:13.951670] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:16:46.092 [2024-11-27 15:21:13.951723] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:46.092 [2024-11-27 15:21:13.951798] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:16:46.092 [2024-11-27 15:21:13.951807] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:46.092 [2024-11-27 15:21:13.951931] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:16:46.092 [2024-11-27 15:21:13.952020] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:16:46.092 [2024-11-27 15:21:13.952031] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006d00 00:16:46.092 [2024-11-27 15:21:13.952109] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:46.092 pt2 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:46.092 15:21:13 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:46.092 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:46.092 "name": "raid_bdev1", 00:16:46.092 "uuid": "cc1e72b4-aa2b-41e9-b157-0ba7845dc144", 00:16:46.092 "strip_size_kb": 0, 00:16:46.092 "state": "online", 00:16:46.092 "raid_level": "raid1", 00:16:46.092 "superblock": true, 00:16:46.092 "num_base_bdevs": 2, 00:16:46.092 "num_base_bdevs_discovered": 1, 00:16:46.092 "num_base_bdevs_operational": 1, 00:16:46.092 "base_bdevs_list": [ 00:16:46.092 { 00:16:46.092 "name": null, 00:16:46.092 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:46.092 "is_configured": false, 00:16:46.092 "data_offset": 256, 00:16:46.092 "data_size": 7936 00:16:46.092 }, 00:16:46.092 { 00:16:46.092 "name": "pt2", 00:16:46.092 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:46.092 "is_configured": true, 00:16:46.092 "data_offset": 256, 00:16:46.092 "data_size": 7936 00:16:46.092 } 00:16:46.092 ] 00:16:46.092 }' 00:16:46.092 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:46.092 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:46.352 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:46.352 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:46.352 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:46.352 [2024-11-27 15:21:14.420663] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:46.352 [2024-11-27 15:21:14.420690] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:46.352 [2024-11-27 15:21:14.420743] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:46.352 [2024-11-27 15:21:14.420786] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:46.352 [2024-11-27 15:21:14.420797] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name raid_bdev1, state offline 00:16:46.352 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:46.352 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:46.352 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:46.352 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:16:46.352 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:46.352 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:46.613 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:16:46.613 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:16:46.613 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:16:46.613 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:46.613 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:46.613 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:46.613 [2024-11-27 15:21:14.484561] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:46.613 [2024-11-27 15:21:14.484661] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:46.613 [2024-11-27 15:21:14.484707] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:16:46.613 [2024-11-27 15:21:14.484744] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:46.613 [2024-11-27 15:21:14.486882] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:46.613 [2024-11-27 15:21:14.486970] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:46.613 [2024-11-27 15:21:14.487041] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:16:46.613 [2024-11-27 15:21:14.487106] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:46.613 [2024-11-27 15:21:14.487209] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:16:46.613 [2024-11-27 15:21:14.487271] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:46.613 [2024-11-27 15:21:14.487336] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007080 name raid_bdev1, state configuring 00:16:46.613 [2024-11-27 15:21:14.487425] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:46.613 [2024-11-27 15:21:14.487548] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007400 00:16:46.613 [2024-11-27 15:21:14.487597] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:46.613 [2024-11-27 15:21:14.487685] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:16:46.613 [2024-11-27 15:21:14.487780] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007400 00:16:46.613 [2024-11-27 15:21:14.487816] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007400 00:16:46.613 [2024-11-27 15:21:14.487956] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:46.613 pt1 00:16:46.613 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:46.613 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:16:46.613 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:46.613 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:46.613 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:46.613 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:46.613 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:46.613 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:46.613 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:46.613 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:46.613 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:46.613 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:46.613 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:46.613 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:46.613 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:46.613 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:46.613 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:46.613 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:46.613 "name": "raid_bdev1", 00:16:46.613 "uuid": "cc1e72b4-aa2b-41e9-b157-0ba7845dc144", 00:16:46.613 "strip_size_kb": 0, 00:16:46.613 "state": "online", 00:16:46.613 "raid_level": "raid1", 00:16:46.613 "superblock": true, 00:16:46.613 "num_base_bdevs": 2, 00:16:46.613 "num_base_bdevs_discovered": 1, 00:16:46.613 "num_base_bdevs_operational": 1, 00:16:46.613 "base_bdevs_list": [ 00:16:46.613 { 00:16:46.613 "name": null, 00:16:46.613 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:46.613 "is_configured": false, 00:16:46.613 "data_offset": 256, 00:16:46.613 "data_size": 7936 00:16:46.613 }, 00:16:46.613 { 00:16:46.613 "name": "pt2", 00:16:46.613 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:46.613 "is_configured": true, 00:16:46.613 "data_offset": 256, 00:16:46.613 "data_size": 7936 00:16:46.613 } 00:16:46.613 ] 00:16:46.613 }' 00:16:46.613 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:46.613 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:46.873 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:16:46.873 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:46.873 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:46.873 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:16:47.133 15:21:14 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:47.133 15:21:15 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:16:47.133 15:21:15 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:47.133 15:21:15 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:47.133 15:21:15 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:47.133 15:21:15 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:16:47.133 [2024-11-27 15:21:15.032162] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:47.133 15:21:15 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:47.133 15:21:15 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # '[' cc1e72b4-aa2b-41e9-b157-0ba7845dc144 '!=' cc1e72b4-aa2b-41e9-b157-0ba7845dc144 ']' 00:16:47.133 15:21:15 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@563 -- # killprocess 99223 00:16:47.133 15:21:15 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@954 -- # '[' -z 99223 ']' 00:16:47.133 15:21:15 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@958 -- # kill -0 99223 00:16:47.133 15:21:15 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@959 -- # uname 00:16:47.133 15:21:15 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:47.134 15:21:15 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 99223 00:16:47.134 15:21:15 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:47.134 15:21:15 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:47.134 15:21:15 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@972 -- # echo 'killing process with pid 99223' 00:16:47.134 killing process with pid 99223 00:16:47.134 15:21:15 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@973 -- # kill 99223 00:16:47.134 [2024-11-27 15:21:15.118656] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:47.134 [2024-11-27 15:21:15.118723] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:47.134 [2024-11-27 15:21:15.118770] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:47.134 [2024-11-27 15:21:15.118779] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007400 name raid_bdev1, state offline 00:16:47.134 15:21:15 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@978 -- # wait 99223 00:16:47.134 [2024-11-27 15:21:15.163283] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:47.704 ************************************ 00:16:47.704 END TEST raid_superblock_test_md_interleaved 00:16:47.704 ************************************ 00:16:47.704 15:21:15 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@565 -- # return 0 00:16:47.704 00:16:47.704 real 0m5.165s 00:16:47.704 user 0m8.320s 00:16:47.704 sys 0m1.144s 00:16:47.704 15:21:15 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:47.704 15:21:15 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:47.704 15:21:15 bdev_raid -- bdev/bdev_raid.sh@1013 -- # run_test raid_rebuild_test_sb_md_interleaved raid_rebuild_test raid1 2 true false false 00:16:47.704 15:21:15 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:16:47.704 15:21:15 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:47.704 15:21:15 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:47.704 ************************************ 00:16:47.704 START TEST raid_rebuild_test_sb_md_interleaved 00:16:47.704 ************************************ 00:16:47.704 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true false false 00:16:47.704 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:16:47.704 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:16:47.704 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:16:47.704 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:16:47.704 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@573 -- # local verify=false 00:16:47.704 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:16:47.704 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:47.704 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:16:47.704 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:47.704 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:47.704 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:16:47.704 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:47.704 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:47.704 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:16:47.704 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:16:47.704 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:16:47.704 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # local strip_size 00:16:47.704 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@577 -- # local create_arg 00:16:47.704 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:16:47.704 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@579 -- # local data_offset 00:16:47.705 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:16:47.705 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:16:47.705 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:16:47.705 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:16:47.705 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@597 -- # raid_pid=99540 00:16:47.705 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:16:47.705 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@598 -- # waitforlisten 99540 00:16:47.705 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:47.705 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@835 -- # '[' -z 99540 ']' 00:16:47.705 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:47.705 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:47.705 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:47.705 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:47.705 15:21:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:47.705 [2024-11-27 15:21:15.690391] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:16:47.705 I/O size of 3145728 is greater than zero copy threshold (65536). 00:16:47.705 Zero copy mechanism will not be used. 00:16:47.705 [2024-11-27 15:21:15.690940] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid99540 ] 00:16:47.965 [2024-11-27 15:21:15.862699] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:47.965 [2024-11-27 15:21:15.902729] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:47.965 [2024-11-27 15:21:15.978977] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:47.965 [2024-11-27 15:21:15.979018] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:48.535 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:48.535 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@868 -- # return 0 00:16:48.535 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:48.535 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev1_malloc 00:16:48.535 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:48.535 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:48.535 BaseBdev1_malloc 00:16:48.535 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:48.535 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:16:48.535 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:48.535 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:48.535 [2024-11-27 15:21:16.536342] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:16:48.535 [2024-11-27 15:21:16.536420] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:48.535 [2024-11-27 15:21:16.536474] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:16:48.535 [2024-11-27 15:21:16.536494] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:48.535 [2024-11-27 15:21:16.538698] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:48.535 [2024-11-27 15:21:16.538739] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:48.535 BaseBdev1 00:16:48.535 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:48.535 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:48.535 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev2_malloc 00:16:48.535 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:48.535 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:48.535 BaseBdev2_malloc 00:16:48.535 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:48.535 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:16:48.535 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:48.535 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:48.535 [2024-11-27 15:21:16.571385] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:16:48.535 [2024-11-27 15:21:16.571453] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:48.535 [2024-11-27 15:21:16.571479] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:16:48.535 [2024-11-27 15:21:16.571492] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:48.535 [2024-11-27 15:21:16.573582] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:48.535 [2024-11-27 15:21:16.573622] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:16:48.535 BaseBdev2 00:16:48.535 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:48.535 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b spare_malloc 00:16:48.535 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:48.535 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:48.535 spare_malloc 00:16:48.535 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:48.535 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:16:48.535 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:48.535 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:48.535 spare_delay 00:16:48.535 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:48.535 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:48.535 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:48.535 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:48.536 [2024-11-27 15:21:16.633331] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:48.536 [2024-11-27 15:21:16.633418] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:48.536 [2024-11-27 15:21:16.633460] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:16:48.536 [2024-11-27 15:21:16.633478] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:48.536 [2024-11-27 15:21:16.636703] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:48.536 [2024-11-27 15:21:16.636762] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:48.796 spare 00:16:48.796 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:48.796 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:16:48.796 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:48.796 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:48.796 [2024-11-27 15:21:16.645397] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:48.796 [2024-11-27 15:21:16.647735] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:48.796 [2024-11-27 15:21:16.648043] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:16:48.796 [2024-11-27 15:21:16.648068] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:48.796 [2024-11-27 15:21:16.648165] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:16:48.796 [2024-11-27 15:21:16.648273] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:16:48.796 [2024-11-27 15:21:16.648289] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:16:48.796 [2024-11-27 15:21:16.648366] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:48.796 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:48.796 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:48.796 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:48.796 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:48.796 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:48.796 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:48.796 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:48.796 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:48.796 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:48.796 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:48.796 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:48.796 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:48.796 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:48.796 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:48.796 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:48.796 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:48.796 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:48.796 "name": "raid_bdev1", 00:16:48.796 "uuid": "bb3fdbb9-1e36-4663-a963-892967f848d4", 00:16:48.796 "strip_size_kb": 0, 00:16:48.796 "state": "online", 00:16:48.796 "raid_level": "raid1", 00:16:48.796 "superblock": true, 00:16:48.796 "num_base_bdevs": 2, 00:16:48.796 "num_base_bdevs_discovered": 2, 00:16:48.796 "num_base_bdevs_operational": 2, 00:16:48.796 "base_bdevs_list": [ 00:16:48.796 { 00:16:48.796 "name": "BaseBdev1", 00:16:48.796 "uuid": "0d147786-ea2a-514a-a844-e715f051cf0e", 00:16:48.796 "is_configured": true, 00:16:48.796 "data_offset": 256, 00:16:48.796 "data_size": 7936 00:16:48.796 }, 00:16:48.796 { 00:16:48.796 "name": "BaseBdev2", 00:16:48.796 "uuid": "0ab27f92-7463-5b57-b042-ca68e865afd0", 00:16:48.796 "is_configured": true, 00:16:48.796 "data_offset": 256, 00:16:48.796 "data_size": 7936 00:16:48.796 } 00:16:48.796 ] 00:16:48.796 }' 00:16:48.796 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:48.796 15:21:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:49.057 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:49.057 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:49.057 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:49.057 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:16:49.057 [2024-11-27 15:21:17.120849] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:49.057 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:49.057 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:16:49.057 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:49.057 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:16:49.057 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:49.057 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:49.317 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:49.317 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:16:49.317 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:16:49.317 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@624 -- # '[' false = true ']' 00:16:49.317 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:16:49.317 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:49.317 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:49.317 [2024-11-27 15:21:17.204421] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:49.317 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:49.317 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:49.317 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:49.317 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:49.317 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:49.317 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:49.317 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:49.317 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:49.317 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:49.317 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:49.317 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:49.317 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:49.317 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:49.317 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:49.317 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:49.317 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:49.317 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:49.317 "name": "raid_bdev1", 00:16:49.317 "uuid": "bb3fdbb9-1e36-4663-a963-892967f848d4", 00:16:49.317 "strip_size_kb": 0, 00:16:49.317 "state": "online", 00:16:49.317 "raid_level": "raid1", 00:16:49.317 "superblock": true, 00:16:49.317 "num_base_bdevs": 2, 00:16:49.317 "num_base_bdevs_discovered": 1, 00:16:49.317 "num_base_bdevs_operational": 1, 00:16:49.317 "base_bdevs_list": [ 00:16:49.317 { 00:16:49.317 "name": null, 00:16:49.317 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:49.317 "is_configured": false, 00:16:49.317 "data_offset": 0, 00:16:49.317 "data_size": 7936 00:16:49.317 }, 00:16:49.317 { 00:16:49.317 "name": "BaseBdev2", 00:16:49.317 "uuid": "0ab27f92-7463-5b57-b042-ca68e865afd0", 00:16:49.317 "is_configured": true, 00:16:49.317 "data_offset": 256, 00:16:49.317 "data_size": 7936 00:16:49.317 } 00:16:49.317 ] 00:16:49.317 }' 00:16:49.317 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:49.317 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:49.577 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:49.577 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:49.577 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:49.577 [2024-11-27 15:21:17.643690] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:49.577 [2024-11-27 15:21:17.650179] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:16:49.577 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:49.577 15:21:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@647 -- # sleep 1 00:16:49.577 [2024-11-27 15:21:17.652356] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:50.960 "name": "raid_bdev1", 00:16:50.960 "uuid": "bb3fdbb9-1e36-4663-a963-892967f848d4", 00:16:50.960 "strip_size_kb": 0, 00:16:50.960 "state": "online", 00:16:50.960 "raid_level": "raid1", 00:16:50.960 "superblock": true, 00:16:50.960 "num_base_bdevs": 2, 00:16:50.960 "num_base_bdevs_discovered": 2, 00:16:50.960 "num_base_bdevs_operational": 2, 00:16:50.960 "process": { 00:16:50.960 "type": "rebuild", 00:16:50.960 "target": "spare", 00:16:50.960 "progress": { 00:16:50.960 "blocks": 2560, 00:16:50.960 "percent": 32 00:16:50.960 } 00:16:50.960 }, 00:16:50.960 "base_bdevs_list": [ 00:16:50.960 { 00:16:50.960 "name": "spare", 00:16:50.960 "uuid": "c6bb4548-a82a-5845-8844-76808e1769b2", 00:16:50.960 "is_configured": true, 00:16:50.960 "data_offset": 256, 00:16:50.960 "data_size": 7936 00:16:50.960 }, 00:16:50.960 { 00:16:50.960 "name": "BaseBdev2", 00:16:50.960 "uuid": "0ab27f92-7463-5b57-b042-ca68e865afd0", 00:16:50.960 "is_configured": true, 00:16:50.960 "data_offset": 256, 00:16:50.960 "data_size": 7936 00:16:50.960 } 00:16:50.960 ] 00:16:50.960 }' 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:50.960 [2024-11-27 15:21:18.813265] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:50.960 [2024-11-27 15:21:18.861165] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:50.960 [2024-11-27 15:21:18.861279] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:50.960 [2024-11-27 15:21:18.861320] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:50.960 [2024-11-27 15:21:18.861345] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:50.960 "name": "raid_bdev1", 00:16:50.960 "uuid": "bb3fdbb9-1e36-4663-a963-892967f848d4", 00:16:50.960 "strip_size_kb": 0, 00:16:50.960 "state": "online", 00:16:50.960 "raid_level": "raid1", 00:16:50.960 "superblock": true, 00:16:50.960 "num_base_bdevs": 2, 00:16:50.960 "num_base_bdevs_discovered": 1, 00:16:50.960 "num_base_bdevs_operational": 1, 00:16:50.960 "base_bdevs_list": [ 00:16:50.960 { 00:16:50.960 "name": null, 00:16:50.960 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:50.960 "is_configured": false, 00:16:50.960 "data_offset": 0, 00:16:50.960 "data_size": 7936 00:16:50.960 }, 00:16:50.960 { 00:16:50.960 "name": "BaseBdev2", 00:16:50.960 "uuid": "0ab27f92-7463-5b57-b042-ca68e865afd0", 00:16:50.960 "is_configured": true, 00:16:50.960 "data_offset": 256, 00:16:50.960 "data_size": 7936 00:16:50.960 } 00:16:50.960 ] 00:16:50.960 }' 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:50.960 15:21:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:51.221 15:21:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:51.221 15:21:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:51.221 15:21:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:51.221 15:21:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:51.221 15:21:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:51.221 15:21:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:51.221 15:21:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:51.221 15:21:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:51.221 15:21:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:51.221 15:21:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:51.221 15:21:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:51.221 "name": "raid_bdev1", 00:16:51.221 "uuid": "bb3fdbb9-1e36-4663-a963-892967f848d4", 00:16:51.221 "strip_size_kb": 0, 00:16:51.221 "state": "online", 00:16:51.221 "raid_level": "raid1", 00:16:51.221 "superblock": true, 00:16:51.221 "num_base_bdevs": 2, 00:16:51.221 "num_base_bdevs_discovered": 1, 00:16:51.221 "num_base_bdevs_operational": 1, 00:16:51.221 "base_bdevs_list": [ 00:16:51.221 { 00:16:51.221 "name": null, 00:16:51.221 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:51.221 "is_configured": false, 00:16:51.221 "data_offset": 0, 00:16:51.221 "data_size": 7936 00:16:51.221 }, 00:16:51.221 { 00:16:51.221 "name": "BaseBdev2", 00:16:51.221 "uuid": "0ab27f92-7463-5b57-b042-ca68e865afd0", 00:16:51.221 "is_configured": true, 00:16:51.221 "data_offset": 256, 00:16:51.221 "data_size": 7936 00:16:51.221 } 00:16:51.221 ] 00:16:51.221 }' 00:16:51.221 15:21:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:51.481 15:21:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:51.481 15:21:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:51.481 15:21:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:51.481 15:21:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:51.481 15:21:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:51.481 15:21:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:51.481 [2024-11-27 15:21:19.419625] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:51.481 [2024-11-27 15:21:19.425326] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:16:51.481 15:21:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:51.481 15:21:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@663 -- # sleep 1 00:16:51.481 [2024-11-27 15:21:19.427456] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:52.421 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:52.421 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:52.421 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:52.421 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:52.421 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:52.421 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:52.421 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:52.421 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:52.421 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:52.421 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:52.421 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:52.421 "name": "raid_bdev1", 00:16:52.421 "uuid": "bb3fdbb9-1e36-4663-a963-892967f848d4", 00:16:52.421 "strip_size_kb": 0, 00:16:52.421 "state": "online", 00:16:52.421 "raid_level": "raid1", 00:16:52.421 "superblock": true, 00:16:52.421 "num_base_bdevs": 2, 00:16:52.421 "num_base_bdevs_discovered": 2, 00:16:52.421 "num_base_bdevs_operational": 2, 00:16:52.421 "process": { 00:16:52.421 "type": "rebuild", 00:16:52.421 "target": "spare", 00:16:52.421 "progress": { 00:16:52.421 "blocks": 2560, 00:16:52.421 "percent": 32 00:16:52.421 } 00:16:52.421 }, 00:16:52.421 "base_bdevs_list": [ 00:16:52.421 { 00:16:52.421 "name": "spare", 00:16:52.421 "uuid": "c6bb4548-a82a-5845-8844-76808e1769b2", 00:16:52.421 "is_configured": true, 00:16:52.421 "data_offset": 256, 00:16:52.421 "data_size": 7936 00:16:52.421 }, 00:16:52.421 { 00:16:52.421 "name": "BaseBdev2", 00:16:52.421 "uuid": "0ab27f92-7463-5b57-b042-ca68e865afd0", 00:16:52.421 "is_configured": true, 00:16:52.421 "data_offset": 256, 00:16:52.421 "data_size": 7936 00:16:52.421 } 00:16:52.421 ] 00:16:52.421 }' 00:16:52.421 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:52.681 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:52.681 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:52.681 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:52.681 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:16:52.681 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:16:52.681 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:16:52.681 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:16:52.681 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:16:52.681 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:16:52.681 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@706 -- # local timeout=621 00:16:52.681 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:52.681 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:52.681 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:52.681 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:52.681 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:52.681 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:52.681 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:52.681 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:52.681 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:52.681 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:52.681 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:52.681 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:52.681 "name": "raid_bdev1", 00:16:52.681 "uuid": "bb3fdbb9-1e36-4663-a963-892967f848d4", 00:16:52.681 "strip_size_kb": 0, 00:16:52.681 "state": "online", 00:16:52.681 "raid_level": "raid1", 00:16:52.681 "superblock": true, 00:16:52.681 "num_base_bdevs": 2, 00:16:52.681 "num_base_bdevs_discovered": 2, 00:16:52.681 "num_base_bdevs_operational": 2, 00:16:52.681 "process": { 00:16:52.681 "type": "rebuild", 00:16:52.681 "target": "spare", 00:16:52.681 "progress": { 00:16:52.681 "blocks": 2816, 00:16:52.681 "percent": 35 00:16:52.681 } 00:16:52.681 }, 00:16:52.681 "base_bdevs_list": [ 00:16:52.681 { 00:16:52.681 "name": "spare", 00:16:52.681 "uuid": "c6bb4548-a82a-5845-8844-76808e1769b2", 00:16:52.681 "is_configured": true, 00:16:52.681 "data_offset": 256, 00:16:52.681 "data_size": 7936 00:16:52.681 }, 00:16:52.681 { 00:16:52.681 "name": "BaseBdev2", 00:16:52.681 "uuid": "0ab27f92-7463-5b57-b042-ca68e865afd0", 00:16:52.681 "is_configured": true, 00:16:52.681 "data_offset": 256, 00:16:52.681 "data_size": 7936 00:16:52.681 } 00:16:52.681 ] 00:16:52.681 }' 00:16:52.681 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:52.681 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:52.681 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:52.681 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:52.681 15:21:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:53.622 15:21:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:53.622 15:21:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:53.622 15:21:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:53.622 15:21:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:53.622 15:21:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:53.622 15:21:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:53.882 15:21:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:53.882 15:21:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:53.882 15:21:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:53.882 15:21:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:53.882 15:21:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:53.882 15:21:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:53.882 "name": "raid_bdev1", 00:16:53.882 "uuid": "bb3fdbb9-1e36-4663-a963-892967f848d4", 00:16:53.882 "strip_size_kb": 0, 00:16:53.882 "state": "online", 00:16:53.882 "raid_level": "raid1", 00:16:53.882 "superblock": true, 00:16:53.882 "num_base_bdevs": 2, 00:16:53.882 "num_base_bdevs_discovered": 2, 00:16:53.882 "num_base_bdevs_operational": 2, 00:16:53.882 "process": { 00:16:53.882 "type": "rebuild", 00:16:53.882 "target": "spare", 00:16:53.882 "progress": { 00:16:53.882 "blocks": 5632, 00:16:53.882 "percent": 70 00:16:53.882 } 00:16:53.882 }, 00:16:53.882 "base_bdevs_list": [ 00:16:53.882 { 00:16:53.882 "name": "spare", 00:16:53.882 "uuid": "c6bb4548-a82a-5845-8844-76808e1769b2", 00:16:53.882 "is_configured": true, 00:16:53.882 "data_offset": 256, 00:16:53.882 "data_size": 7936 00:16:53.882 }, 00:16:53.882 { 00:16:53.882 "name": "BaseBdev2", 00:16:53.882 "uuid": "0ab27f92-7463-5b57-b042-ca68e865afd0", 00:16:53.882 "is_configured": true, 00:16:53.882 "data_offset": 256, 00:16:53.882 "data_size": 7936 00:16:53.882 } 00:16:53.882 ] 00:16:53.882 }' 00:16:53.882 15:21:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:53.882 15:21:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:53.882 15:21:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:53.882 15:21:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:53.882 15:21:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:54.453 [2024-11-27 15:21:22.548018] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:16:54.453 [2024-11-27 15:21:22.548109] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:16:54.453 [2024-11-27 15:21:22.548222] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:55.024 15:21:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:55.024 15:21:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:55.024 15:21:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:55.024 15:21:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:55.024 15:21:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:55.024 15:21:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:55.024 15:21:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:55.024 15:21:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:55.024 15:21:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:55.024 15:21:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:55.024 15:21:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:55.024 15:21:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:55.024 "name": "raid_bdev1", 00:16:55.024 "uuid": "bb3fdbb9-1e36-4663-a963-892967f848d4", 00:16:55.024 "strip_size_kb": 0, 00:16:55.024 "state": "online", 00:16:55.024 "raid_level": "raid1", 00:16:55.024 "superblock": true, 00:16:55.024 "num_base_bdevs": 2, 00:16:55.024 "num_base_bdevs_discovered": 2, 00:16:55.024 "num_base_bdevs_operational": 2, 00:16:55.024 "base_bdevs_list": [ 00:16:55.024 { 00:16:55.024 "name": "spare", 00:16:55.024 "uuid": "c6bb4548-a82a-5845-8844-76808e1769b2", 00:16:55.024 "is_configured": true, 00:16:55.024 "data_offset": 256, 00:16:55.024 "data_size": 7936 00:16:55.024 }, 00:16:55.024 { 00:16:55.024 "name": "BaseBdev2", 00:16:55.024 "uuid": "0ab27f92-7463-5b57-b042-ca68e865afd0", 00:16:55.024 "is_configured": true, 00:16:55.024 "data_offset": 256, 00:16:55.024 "data_size": 7936 00:16:55.024 } 00:16:55.024 ] 00:16:55.024 }' 00:16:55.024 15:21:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:55.024 15:21:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:16:55.024 15:21:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:55.024 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:16:55.024 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@709 -- # break 00:16:55.024 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:55.024 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:55.024 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:55.024 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:55.025 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:55.025 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:55.025 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:55.025 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:55.025 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:55.025 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:55.025 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:55.025 "name": "raid_bdev1", 00:16:55.025 "uuid": "bb3fdbb9-1e36-4663-a963-892967f848d4", 00:16:55.025 "strip_size_kb": 0, 00:16:55.025 "state": "online", 00:16:55.025 "raid_level": "raid1", 00:16:55.025 "superblock": true, 00:16:55.025 "num_base_bdevs": 2, 00:16:55.025 "num_base_bdevs_discovered": 2, 00:16:55.025 "num_base_bdevs_operational": 2, 00:16:55.025 "base_bdevs_list": [ 00:16:55.025 { 00:16:55.025 "name": "spare", 00:16:55.025 "uuid": "c6bb4548-a82a-5845-8844-76808e1769b2", 00:16:55.025 "is_configured": true, 00:16:55.025 "data_offset": 256, 00:16:55.025 "data_size": 7936 00:16:55.025 }, 00:16:55.025 { 00:16:55.025 "name": "BaseBdev2", 00:16:55.025 "uuid": "0ab27f92-7463-5b57-b042-ca68e865afd0", 00:16:55.025 "is_configured": true, 00:16:55.025 "data_offset": 256, 00:16:55.025 "data_size": 7936 00:16:55.025 } 00:16:55.025 ] 00:16:55.025 }' 00:16:55.025 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:55.025 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:55.025 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:55.285 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:55.285 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:55.285 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:55.285 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:55.285 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:55.285 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:55.285 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:55.285 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:55.285 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:55.285 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:55.285 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:55.285 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:55.286 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:55.286 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:55.286 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:55.286 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:55.286 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:55.286 "name": "raid_bdev1", 00:16:55.286 "uuid": "bb3fdbb9-1e36-4663-a963-892967f848d4", 00:16:55.286 "strip_size_kb": 0, 00:16:55.286 "state": "online", 00:16:55.286 "raid_level": "raid1", 00:16:55.286 "superblock": true, 00:16:55.286 "num_base_bdevs": 2, 00:16:55.286 "num_base_bdevs_discovered": 2, 00:16:55.286 "num_base_bdevs_operational": 2, 00:16:55.286 "base_bdevs_list": [ 00:16:55.286 { 00:16:55.286 "name": "spare", 00:16:55.286 "uuid": "c6bb4548-a82a-5845-8844-76808e1769b2", 00:16:55.286 "is_configured": true, 00:16:55.286 "data_offset": 256, 00:16:55.286 "data_size": 7936 00:16:55.286 }, 00:16:55.286 { 00:16:55.286 "name": "BaseBdev2", 00:16:55.286 "uuid": "0ab27f92-7463-5b57-b042-ca68e865afd0", 00:16:55.286 "is_configured": true, 00:16:55.286 "data_offset": 256, 00:16:55.286 "data_size": 7936 00:16:55.286 } 00:16:55.286 ] 00:16:55.286 }' 00:16:55.286 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:55.286 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:55.546 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:55.546 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:55.546 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:55.546 [2024-11-27 15:21:23.604875] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:55.546 [2024-11-27 15:21:23.604969] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:55.546 [2024-11-27 15:21:23.605095] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:55.546 [2024-11-27 15:21:23.605190] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:55.546 [2024-11-27 15:21:23.605281] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:16:55.546 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:55.546 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:55.546 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:55.546 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:55.546 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # jq length 00:16:55.546 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:55.546 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:16:55.546 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@722 -- # '[' false = true ']' 00:16:55.546 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:16:55.546 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:16:55.546 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:55.546 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:55.807 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:55.807 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:55.807 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:55.807 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:55.807 [2024-11-27 15:21:23.660786] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:55.807 [2024-11-27 15:21:23.660918] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:55.807 [2024-11-27 15:21:23.660980] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:16:55.807 [2024-11-27 15:21:23.661035] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:55.807 [2024-11-27 15:21:23.663342] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:55.807 [2024-11-27 15:21:23.663427] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:55.807 [2024-11-27 15:21:23.663517] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:16:55.807 [2024-11-27 15:21:23.663610] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:55.807 [2024-11-27 15:21:23.663762] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:55.807 spare 00:16:55.807 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:55.807 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:16:55.807 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:55.807 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:55.807 [2024-11-27 15:21:23.763698] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006600 00:16:55.807 [2024-11-27 15:21:23.763725] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:55.807 [2024-11-27 15:21:23.763855] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:16:55.807 [2024-11-27 15:21:23.763957] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006600 00:16:55.807 [2024-11-27 15:21:23.763972] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006600 00:16:55.807 [2024-11-27 15:21:23.764045] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:55.807 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:55.807 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:55.807 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:55.807 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:55.807 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:55.807 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:55.808 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:55.808 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:55.808 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:55.808 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:55.808 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:55.808 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:55.808 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:55.808 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:55.808 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:55.808 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:55.808 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:55.808 "name": "raid_bdev1", 00:16:55.808 "uuid": "bb3fdbb9-1e36-4663-a963-892967f848d4", 00:16:55.808 "strip_size_kb": 0, 00:16:55.808 "state": "online", 00:16:55.808 "raid_level": "raid1", 00:16:55.808 "superblock": true, 00:16:55.808 "num_base_bdevs": 2, 00:16:55.808 "num_base_bdevs_discovered": 2, 00:16:55.808 "num_base_bdevs_operational": 2, 00:16:55.808 "base_bdevs_list": [ 00:16:55.808 { 00:16:55.808 "name": "spare", 00:16:55.808 "uuid": "c6bb4548-a82a-5845-8844-76808e1769b2", 00:16:55.808 "is_configured": true, 00:16:55.808 "data_offset": 256, 00:16:55.808 "data_size": 7936 00:16:55.808 }, 00:16:55.808 { 00:16:55.808 "name": "BaseBdev2", 00:16:55.808 "uuid": "0ab27f92-7463-5b57-b042-ca68e865afd0", 00:16:55.808 "is_configured": true, 00:16:55.808 "data_offset": 256, 00:16:55.808 "data_size": 7936 00:16:55.808 } 00:16:55.808 ] 00:16:55.808 }' 00:16:55.808 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:55.808 15:21:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:56.068 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:56.068 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:56.068 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:56.068 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:56.068 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:56.068 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:56.068 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:56.068 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:56.068 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:56.068 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:56.328 "name": "raid_bdev1", 00:16:56.328 "uuid": "bb3fdbb9-1e36-4663-a963-892967f848d4", 00:16:56.328 "strip_size_kb": 0, 00:16:56.328 "state": "online", 00:16:56.328 "raid_level": "raid1", 00:16:56.328 "superblock": true, 00:16:56.328 "num_base_bdevs": 2, 00:16:56.328 "num_base_bdevs_discovered": 2, 00:16:56.328 "num_base_bdevs_operational": 2, 00:16:56.328 "base_bdevs_list": [ 00:16:56.328 { 00:16:56.328 "name": "spare", 00:16:56.328 "uuid": "c6bb4548-a82a-5845-8844-76808e1769b2", 00:16:56.328 "is_configured": true, 00:16:56.328 "data_offset": 256, 00:16:56.328 "data_size": 7936 00:16:56.328 }, 00:16:56.328 { 00:16:56.328 "name": "BaseBdev2", 00:16:56.328 "uuid": "0ab27f92-7463-5b57-b042-ca68e865afd0", 00:16:56.328 "is_configured": true, 00:16:56.328 "data_offset": 256, 00:16:56.328 "data_size": 7936 00:16:56.328 } 00:16:56.328 ] 00:16:56.328 }' 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:56.328 [2024-11-27 15:21:24.347668] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:56.328 "name": "raid_bdev1", 00:16:56.328 "uuid": "bb3fdbb9-1e36-4663-a963-892967f848d4", 00:16:56.328 "strip_size_kb": 0, 00:16:56.328 "state": "online", 00:16:56.328 "raid_level": "raid1", 00:16:56.328 "superblock": true, 00:16:56.328 "num_base_bdevs": 2, 00:16:56.328 "num_base_bdevs_discovered": 1, 00:16:56.328 "num_base_bdevs_operational": 1, 00:16:56.328 "base_bdevs_list": [ 00:16:56.328 { 00:16:56.328 "name": null, 00:16:56.328 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:56.328 "is_configured": false, 00:16:56.328 "data_offset": 0, 00:16:56.328 "data_size": 7936 00:16:56.328 }, 00:16:56.328 { 00:16:56.328 "name": "BaseBdev2", 00:16:56.328 "uuid": "0ab27f92-7463-5b57-b042-ca68e865afd0", 00:16:56.328 "is_configured": true, 00:16:56.328 "data_offset": 256, 00:16:56.328 "data_size": 7936 00:16:56.328 } 00:16:56.328 ] 00:16:56.328 }' 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:56.328 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:56.898 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:56.898 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:56.898 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:56.898 [2024-11-27 15:21:24.759204] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:56.898 [2024-11-27 15:21:24.759395] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:16:56.898 [2024-11-27 15:21:24.759468] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:16:56.898 [2024-11-27 15:21:24.759534] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:56.898 [2024-11-27 15:21:24.765376] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:16:56.898 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:56.898 15:21:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@757 -- # sleep 1 00:16:56.898 [2024-11-27 15:21:24.767402] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:57.837 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:57.837 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:57.837 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:57.837 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:57.837 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:57.837 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:57.837 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:57.837 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:57.837 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:57.837 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:57.837 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:57.837 "name": "raid_bdev1", 00:16:57.837 "uuid": "bb3fdbb9-1e36-4663-a963-892967f848d4", 00:16:57.837 "strip_size_kb": 0, 00:16:57.837 "state": "online", 00:16:57.837 "raid_level": "raid1", 00:16:57.837 "superblock": true, 00:16:57.837 "num_base_bdevs": 2, 00:16:57.837 "num_base_bdevs_discovered": 2, 00:16:57.837 "num_base_bdevs_operational": 2, 00:16:57.837 "process": { 00:16:57.837 "type": "rebuild", 00:16:57.837 "target": "spare", 00:16:57.837 "progress": { 00:16:57.837 "blocks": 2560, 00:16:57.837 "percent": 32 00:16:57.837 } 00:16:57.837 }, 00:16:57.837 "base_bdevs_list": [ 00:16:57.837 { 00:16:57.837 "name": "spare", 00:16:57.837 "uuid": "c6bb4548-a82a-5845-8844-76808e1769b2", 00:16:57.837 "is_configured": true, 00:16:57.837 "data_offset": 256, 00:16:57.837 "data_size": 7936 00:16:57.837 }, 00:16:57.837 { 00:16:57.837 "name": "BaseBdev2", 00:16:57.837 "uuid": "0ab27f92-7463-5b57-b042-ca68e865afd0", 00:16:57.837 "is_configured": true, 00:16:57.837 "data_offset": 256, 00:16:57.837 "data_size": 7936 00:16:57.837 } 00:16:57.837 ] 00:16:57.837 }' 00:16:57.837 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:57.837 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:57.837 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:57.837 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:57.837 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:16:57.837 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:57.837 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:57.837 [2024-11-27 15:21:25.931643] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:58.096 [2024-11-27 15:21:25.975256] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:58.096 [2024-11-27 15:21:25.975317] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:58.096 [2024-11-27 15:21:25.975337] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:58.096 [2024-11-27 15:21:25.975346] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:58.096 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:58.096 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:58.096 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:58.096 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:58.096 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:58.096 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:58.096 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:58.097 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:58.097 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:58.097 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:58.097 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:58.097 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:58.097 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:58.097 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:58.097 15:21:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:58.097 15:21:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:58.097 15:21:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:58.097 "name": "raid_bdev1", 00:16:58.097 "uuid": "bb3fdbb9-1e36-4663-a963-892967f848d4", 00:16:58.097 "strip_size_kb": 0, 00:16:58.097 "state": "online", 00:16:58.097 "raid_level": "raid1", 00:16:58.097 "superblock": true, 00:16:58.097 "num_base_bdevs": 2, 00:16:58.097 "num_base_bdevs_discovered": 1, 00:16:58.097 "num_base_bdevs_operational": 1, 00:16:58.097 "base_bdevs_list": [ 00:16:58.097 { 00:16:58.097 "name": null, 00:16:58.097 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:58.097 "is_configured": false, 00:16:58.097 "data_offset": 0, 00:16:58.097 "data_size": 7936 00:16:58.097 }, 00:16:58.097 { 00:16:58.097 "name": "BaseBdev2", 00:16:58.097 "uuid": "0ab27f92-7463-5b57-b042-ca68e865afd0", 00:16:58.097 "is_configured": true, 00:16:58.097 "data_offset": 256, 00:16:58.097 "data_size": 7936 00:16:58.097 } 00:16:58.097 ] 00:16:58.097 }' 00:16:58.097 15:21:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:58.097 15:21:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:58.356 15:21:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:58.356 15:21:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:58.356 15:21:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:58.356 [2024-11-27 15:21:26.396679] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:58.356 [2024-11-27 15:21:26.396800] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:58.356 [2024-11-27 15:21:26.396853] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:16:58.356 [2024-11-27 15:21:26.396886] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:58.356 [2024-11-27 15:21:26.397135] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:58.356 [2024-11-27 15:21:26.397190] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:58.356 [2024-11-27 15:21:26.397282] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:16:58.356 [2024-11-27 15:21:26.397324] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:16:58.356 [2024-11-27 15:21:26.397379] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:16:58.356 [2024-11-27 15:21:26.397453] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:58.356 [2024-11-27 15:21:26.401758] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:16:58.356 spare 00:16:58.356 15:21:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:58.356 15:21:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@764 -- # sleep 1 00:16:58.356 [2024-11-27 15:21:26.403941] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:59.739 "name": "raid_bdev1", 00:16:59.739 "uuid": "bb3fdbb9-1e36-4663-a963-892967f848d4", 00:16:59.739 "strip_size_kb": 0, 00:16:59.739 "state": "online", 00:16:59.739 "raid_level": "raid1", 00:16:59.739 "superblock": true, 00:16:59.739 "num_base_bdevs": 2, 00:16:59.739 "num_base_bdevs_discovered": 2, 00:16:59.739 "num_base_bdevs_operational": 2, 00:16:59.739 "process": { 00:16:59.739 "type": "rebuild", 00:16:59.739 "target": "spare", 00:16:59.739 "progress": { 00:16:59.739 "blocks": 2560, 00:16:59.739 "percent": 32 00:16:59.739 } 00:16:59.739 }, 00:16:59.739 "base_bdevs_list": [ 00:16:59.739 { 00:16:59.739 "name": "spare", 00:16:59.739 "uuid": "c6bb4548-a82a-5845-8844-76808e1769b2", 00:16:59.739 "is_configured": true, 00:16:59.739 "data_offset": 256, 00:16:59.739 "data_size": 7936 00:16:59.739 }, 00:16:59.739 { 00:16:59.739 "name": "BaseBdev2", 00:16:59.739 "uuid": "0ab27f92-7463-5b57-b042-ca68e865afd0", 00:16:59.739 "is_configured": true, 00:16:59.739 "data_offset": 256, 00:16:59.739 "data_size": 7936 00:16:59.739 } 00:16:59.739 ] 00:16:59.739 }' 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:59.739 [2024-11-27 15:21:27.560064] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:59.739 [2024-11-27 15:21:27.611634] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:59.739 [2024-11-27 15:21:27.611705] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:59.739 [2024-11-27 15:21:27.611722] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:59.739 [2024-11-27 15:21:27.611734] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:59.739 "name": "raid_bdev1", 00:16:59.739 "uuid": "bb3fdbb9-1e36-4663-a963-892967f848d4", 00:16:59.739 "strip_size_kb": 0, 00:16:59.739 "state": "online", 00:16:59.739 "raid_level": "raid1", 00:16:59.739 "superblock": true, 00:16:59.739 "num_base_bdevs": 2, 00:16:59.739 "num_base_bdevs_discovered": 1, 00:16:59.739 "num_base_bdevs_operational": 1, 00:16:59.739 "base_bdevs_list": [ 00:16:59.739 { 00:16:59.739 "name": null, 00:16:59.739 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:59.739 "is_configured": false, 00:16:59.739 "data_offset": 0, 00:16:59.739 "data_size": 7936 00:16:59.739 }, 00:16:59.739 { 00:16:59.739 "name": "BaseBdev2", 00:16:59.739 "uuid": "0ab27f92-7463-5b57-b042-ca68e865afd0", 00:16:59.739 "is_configured": true, 00:16:59.739 "data_offset": 256, 00:16:59.739 "data_size": 7936 00:16:59.739 } 00:16:59.739 ] 00:16:59.739 }' 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:59.739 15:21:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:00.000 15:21:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:00.000 15:21:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:00.000 15:21:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:00.000 15:21:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:00.000 15:21:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:00.000 15:21:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:00.000 15:21:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:00.000 15:21:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:00.000 15:21:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:00.000 15:21:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:00.260 15:21:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:00.260 "name": "raid_bdev1", 00:17:00.260 "uuid": "bb3fdbb9-1e36-4663-a963-892967f848d4", 00:17:00.260 "strip_size_kb": 0, 00:17:00.260 "state": "online", 00:17:00.260 "raid_level": "raid1", 00:17:00.260 "superblock": true, 00:17:00.260 "num_base_bdevs": 2, 00:17:00.260 "num_base_bdevs_discovered": 1, 00:17:00.260 "num_base_bdevs_operational": 1, 00:17:00.260 "base_bdevs_list": [ 00:17:00.260 { 00:17:00.260 "name": null, 00:17:00.260 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:00.260 "is_configured": false, 00:17:00.260 "data_offset": 0, 00:17:00.260 "data_size": 7936 00:17:00.260 }, 00:17:00.260 { 00:17:00.260 "name": "BaseBdev2", 00:17:00.260 "uuid": "0ab27f92-7463-5b57-b042-ca68e865afd0", 00:17:00.260 "is_configured": true, 00:17:00.260 "data_offset": 256, 00:17:00.260 "data_size": 7936 00:17:00.260 } 00:17:00.260 ] 00:17:00.260 }' 00:17:00.260 15:21:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:00.260 15:21:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:00.260 15:21:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:00.260 15:21:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:00.260 15:21:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:17:00.260 15:21:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:00.260 15:21:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:00.260 15:21:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:00.260 15:21:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:17:00.260 15:21:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:00.260 15:21:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:00.260 [2024-11-27 15:21:28.216826] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:17:00.260 [2024-11-27 15:21:28.216962] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:00.260 [2024-11-27 15:21:28.216994] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:17:00.261 [2024-11-27 15:21:28.217008] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:00.261 [2024-11-27 15:21:28.217198] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:00.261 [2024-11-27 15:21:28.217218] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:17:00.261 [2024-11-27 15:21:28.217272] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:17:00.261 [2024-11-27 15:21:28.217290] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:17:00.261 [2024-11-27 15:21:28.217300] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:17:00.261 [2024-11-27 15:21:28.217329] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:17:00.261 BaseBdev1 00:17:00.261 15:21:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:00.261 15:21:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@775 -- # sleep 1 00:17:01.202 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:01.202 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:01.202 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:01.202 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:01.202 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:01.202 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:01.202 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:01.202 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:01.202 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:01.202 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:01.202 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:01.202 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:01.202 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:01.202 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:01.202 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:01.203 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:01.203 "name": "raid_bdev1", 00:17:01.203 "uuid": "bb3fdbb9-1e36-4663-a963-892967f848d4", 00:17:01.203 "strip_size_kb": 0, 00:17:01.203 "state": "online", 00:17:01.203 "raid_level": "raid1", 00:17:01.203 "superblock": true, 00:17:01.203 "num_base_bdevs": 2, 00:17:01.203 "num_base_bdevs_discovered": 1, 00:17:01.203 "num_base_bdevs_operational": 1, 00:17:01.203 "base_bdevs_list": [ 00:17:01.203 { 00:17:01.203 "name": null, 00:17:01.203 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:01.203 "is_configured": false, 00:17:01.203 "data_offset": 0, 00:17:01.203 "data_size": 7936 00:17:01.203 }, 00:17:01.203 { 00:17:01.203 "name": "BaseBdev2", 00:17:01.203 "uuid": "0ab27f92-7463-5b57-b042-ca68e865afd0", 00:17:01.203 "is_configured": true, 00:17:01.203 "data_offset": 256, 00:17:01.203 "data_size": 7936 00:17:01.203 } 00:17:01.203 ] 00:17:01.203 }' 00:17:01.203 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:01.203 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:01.815 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:01.815 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:01.815 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:01.815 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:01.815 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:01.815 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:01.815 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:01.815 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:01.815 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:01.815 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:01.815 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:01.815 "name": "raid_bdev1", 00:17:01.816 "uuid": "bb3fdbb9-1e36-4663-a963-892967f848d4", 00:17:01.816 "strip_size_kb": 0, 00:17:01.816 "state": "online", 00:17:01.816 "raid_level": "raid1", 00:17:01.816 "superblock": true, 00:17:01.816 "num_base_bdevs": 2, 00:17:01.816 "num_base_bdevs_discovered": 1, 00:17:01.816 "num_base_bdevs_operational": 1, 00:17:01.816 "base_bdevs_list": [ 00:17:01.816 { 00:17:01.816 "name": null, 00:17:01.816 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:01.816 "is_configured": false, 00:17:01.816 "data_offset": 0, 00:17:01.816 "data_size": 7936 00:17:01.816 }, 00:17:01.816 { 00:17:01.816 "name": "BaseBdev2", 00:17:01.816 "uuid": "0ab27f92-7463-5b57-b042-ca68e865afd0", 00:17:01.816 "is_configured": true, 00:17:01.816 "data_offset": 256, 00:17:01.816 "data_size": 7936 00:17:01.816 } 00:17:01.816 ] 00:17:01.816 }' 00:17:01.816 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:01.816 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:01.816 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:01.816 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:01.816 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:17:01.816 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@652 -- # local es=0 00:17:01.816 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:17:01.816 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:17:01.816 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:17:01.816 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:17:01.816 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:17:01.816 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:17:01.816 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:01.816 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:01.816 [2024-11-27 15:21:29.818124] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:17:01.816 [2024-11-27 15:21:29.818274] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:17:01.816 [2024-11-27 15:21:29.818288] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:17:01.816 request: 00:17:01.816 { 00:17:01.816 "base_bdev": "BaseBdev1", 00:17:01.816 "raid_bdev": "raid_bdev1", 00:17:01.816 "method": "bdev_raid_add_base_bdev", 00:17:01.816 "req_id": 1 00:17:01.816 } 00:17:01.816 Got JSON-RPC error response 00:17:01.816 response: 00:17:01.816 { 00:17:01.816 "code": -22, 00:17:01.816 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:17:01.816 } 00:17:01.816 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:17:01.816 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@655 -- # es=1 00:17:01.816 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:17:01.816 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:17:01.816 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:17:01.816 15:21:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@779 -- # sleep 1 00:17:02.771 15:21:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:02.771 15:21:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:02.771 15:21:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:02.771 15:21:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:02.771 15:21:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:02.771 15:21:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:02.771 15:21:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:02.771 15:21:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:02.771 15:21:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:02.771 15:21:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:02.771 15:21:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:02.771 15:21:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:02.771 15:21:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:02.771 15:21:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:02.771 15:21:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:03.032 15:21:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:03.032 "name": "raid_bdev1", 00:17:03.032 "uuid": "bb3fdbb9-1e36-4663-a963-892967f848d4", 00:17:03.032 "strip_size_kb": 0, 00:17:03.032 "state": "online", 00:17:03.032 "raid_level": "raid1", 00:17:03.032 "superblock": true, 00:17:03.032 "num_base_bdevs": 2, 00:17:03.032 "num_base_bdevs_discovered": 1, 00:17:03.032 "num_base_bdevs_operational": 1, 00:17:03.032 "base_bdevs_list": [ 00:17:03.032 { 00:17:03.032 "name": null, 00:17:03.032 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:03.032 "is_configured": false, 00:17:03.032 "data_offset": 0, 00:17:03.032 "data_size": 7936 00:17:03.032 }, 00:17:03.032 { 00:17:03.032 "name": "BaseBdev2", 00:17:03.032 "uuid": "0ab27f92-7463-5b57-b042-ca68e865afd0", 00:17:03.032 "is_configured": true, 00:17:03.032 "data_offset": 256, 00:17:03.032 "data_size": 7936 00:17:03.032 } 00:17:03.032 ] 00:17:03.032 }' 00:17:03.032 15:21:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:03.032 15:21:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:03.293 15:21:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:03.293 15:21:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:03.293 15:21:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:03.293 15:21:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:03.293 15:21:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:03.293 15:21:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:03.293 15:21:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:03.293 15:21:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:03.293 15:21:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:03.293 15:21:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:03.293 15:21:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:03.293 "name": "raid_bdev1", 00:17:03.293 "uuid": "bb3fdbb9-1e36-4663-a963-892967f848d4", 00:17:03.293 "strip_size_kb": 0, 00:17:03.293 "state": "online", 00:17:03.293 "raid_level": "raid1", 00:17:03.293 "superblock": true, 00:17:03.293 "num_base_bdevs": 2, 00:17:03.293 "num_base_bdevs_discovered": 1, 00:17:03.293 "num_base_bdevs_operational": 1, 00:17:03.293 "base_bdevs_list": [ 00:17:03.293 { 00:17:03.293 "name": null, 00:17:03.293 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:03.293 "is_configured": false, 00:17:03.293 "data_offset": 0, 00:17:03.293 "data_size": 7936 00:17:03.293 }, 00:17:03.293 { 00:17:03.293 "name": "BaseBdev2", 00:17:03.293 "uuid": "0ab27f92-7463-5b57-b042-ca68e865afd0", 00:17:03.293 "is_configured": true, 00:17:03.293 "data_offset": 256, 00:17:03.293 "data_size": 7936 00:17:03.293 } 00:17:03.293 ] 00:17:03.293 }' 00:17:03.293 15:21:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:03.293 15:21:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:03.293 15:21:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:03.554 15:21:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:03.554 15:21:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@784 -- # killprocess 99540 00:17:03.554 15:21:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@954 -- # '[' -z 99540 ']' 00:17:03.554 15:21:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@958 -- # kill -0 99540 00:17:03.554 15:21:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@959 -- # uname 00:17:03.554 15:21:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:17:03.554 15:21:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 99540 00:17:03.554 15:21:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:17:03.554 15:21:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:17:03.554 15:21:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@972 -- # echo 'killing process with pid 99540' 00:17:03.554 killing process with pid 99540 00:17:03.554 Received shutdown signal, test time was about 60.000000 seconds 00:17:03.554 00:17:03.554 Latency(us) 00:17:03.554 [2024-11-27T15:21:31.661Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:17:03.554 [2024-11-27T15:21:31.661Z] =================================================================================================================== 00:17:03.554 [2024-11-27T15:21:31.661Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:17:03.554 15:21:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@973 -- # kill 99540 00:17:03.554 [2024-11-27 15:21:31.455097] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:17:03.554 [2024-11-27 15:21:31.455202] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:03.554 [2024-11-27 15:21:31.455250] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:03.554 [2024-11-27 15:21:31.455260] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state offline 00:17:03.554 15:21:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@978 -- # wait 99540 00:17:03.554 [2024-11-27 15:21:31.518173] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:17:03.814 15:21:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@786 -- # return 0 00:17:03.814 00:17:03.814 real 0m16.261s 00:17:03.814 user 0m21.538s 00:17:03.814 sys 0m1.743s 00:17:03.814 15:21:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:03.814 ************************************ 00:17:03.814 END TEST raid_rebuild_test_sb_md_interleaved 00:17:03.814 ************************************ 00:17:03.814 15:21:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:03.814 15:21:31 bdev_raid -- bdev/bdev_raid.sh@1015 -- # trap - EXIT 00:17:03.814 15:21:31 bdev_raid -- bdev/bdev_raid.sh@1016 -- # cleanup 00:17:03.814 15:21:31 bdev_raid -- bdev/bdev_raid.sh@56 -- # '[' -n 99540 ']' 00:17:03.814 15:21:31 bdev_raid -- bdev/bdev_raid.sh@56 -- # ps -p 99540 00:17:04.073 15:21:31 bdev_raid -- bdev/bdev_raid.sh@60 -- # rm -rf /raidtest 00:17:04.073 ************************************ 00:17:04.073 END TEST bdev_raid 00:17:04.073 ************************************ 00:17:04.073 00:17:04.073 real 10m2.646s 00:17:04.073 user 14m12.272s 00:17:04.073 sys 1m52.452s 00:17:04.073 15:21:31 bdev_raid -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:04.073 15:21:31 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:17:04.073 15:21:32 -- spdk/autotest.sh@190 -- # run_test spdkcli_raid /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:17:04.073 15:21:32 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:17:04.073 15:21:32 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:04.073 15:21:32 -- common/autotest_common.sh@10 -- # set +x 00:17:04.073 ************************************ 00:17:04.073 START TEST spdkcli_raid 00:17:04.073 ************************************ 00:17:04.073 15:21:32 spdkcli_raid -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:17:04.074 * Looking for test storage... 00:17:04.074 * Found test storage at /home/vagrant/spdk_repo/spdk/test/spdkcli 00:17:04.074 15:21:32 spdkcli_raid -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:17:04.074 15:21:32 spdkcli_raid -- common/autotest_common.sh@1693 -- # lcov --version 00:17:04.074 15:21:32 spdkcli_raid -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:17:04.335 15:21:32 spdkcli_raid -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:17:04.335 15:21:32 spdkcli_raid -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:17:04.335 15:21:32 spdkcli_raid -- scripts/common.sh@333 -- # local ver1 ver1_l 00:17:04.335 15:21:32 spdkcli_raid -- scripts/common.sh@334 -- # local ver2 ver2_l 00:17:04.335 15:21:32 spdkcli_raid -- scripts/common.sh@336 -- # IFS=.-: 00:17:04.335 15:21:32 spdkcli_raid -- scripts/common.sh@336 -- # read -ra ver1 00:17:04.335 15:21:32 spdkcli_raid -- scripts/common.sh@337 -- # IFS=.-: 00:17:04.335 15:21:32 spdkcli_raid -- scripts/common.sh@337 -- # read -ra ver2 00:17:04.335 15:21:32 spdkcli_raid -- scripts/common.sh@338 -- # local 'op=<' 00:17:04.335 15:21:32 spdkcli_raid -- scripts/common.sh@340 -- # ver1_l=2 00:17:04.335 15:21:32 spdkcli_raid -- scripts/common.sh@341 -- # ver2_l=1 00:17:04.335 15:21:32 spdkcli_raid -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:17:04.335 15:21:32 spdkcli_raid -- scripts/common.sh@344 -- # case "$op" in 00:17:04.335 15:21:32 spdkcli_raid -- scripts/common.sh@345 -- # : 1 00:17:04.335 15:21:32 spdkcli_raid -- scripts/common.sh@364 -- # (( v = 0 )) 00:17:04.335 15:21:32 spdkcli_raid -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:17:04.335 15:21:32 spdkcli_raid -- scripts/common.sh@365 -- # decimal 1 00:17:04.335 15:21:32 spdkcli_raid -- scripts/common.sh@353 -- # local d=1 00:17:04.335 15:21:32 spdkcli_raid -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:17:04.335 15:21:32 spdkcli_raid -- scripts/common.sh@355 -- # echo 1 00:17:04.335 15:21:32 spdkcli_raid -- scripts/common.sh@365 -- # ver1[v]=1 00:17:04.335 15:21:32 spdkcli_raid -- scripts/common.sh@366 -- # decimal 2 00:17:04.335 15:21:32 spdkcli_raid -- scripts/common.sh@353 -- # local d=2 00:17:04.335 15:21:32 spdkcli_raid -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:17:04.335 15:21:32 spdkcli_raid -- scripts/common.sh@355 -- # echo 2 00:17:04.335 15:21:32 spdkcli_raid -- scripts/common.sh@366 -- # ver2[v]=2 00:17:04.335 15:21:32 spdkcli_raid -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:17:04.335 15:21:32 spdkcli_raid -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:17:04.335 15:21:32 spdkcli_raid -- scripts/common.sh@368 -- # return 0 00:17:04.335 15:21:32 spdkcli_raid -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:17:04.335 15:21:32 spdkcli_raid -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:17:04.335 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:04.335 --rc genhtml_branch_coverage=1 00:17:04.335 --rc genhtml_function_coverage=1 00:17:04.335 --rc genhtml_legend=1 00:17:04.335 --rc geninfo_all_blocks=1 00:17:04.335 --rc geninfo_unexecuted_blocks=1 00:17:04.335 00:17:04.335 ' 00:17:04.335 15:21:32 spdkcli_raid -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:17:04.335 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:04.335 --rc genhtml_branch_coverage=1 00:17:04.335 --rc genhtml_function_coverage=1 00:17:04.335 --rc genhtml_legend=1 00:17:04.335 --rc geninfo_all_blocks=1 00:17:04.335 --rc geninfo_unexecuted_blocks=1 00:17:04.335 00:17:04.335 ' 00:17:04.335 15:21:32 spdkcli_raid -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:17:04.335 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:04.335 --rc genhtml_branch_coverage=1 00:17:04.335 --rc genhtml_function_coverage=1 00:17:04.335 --rc genhtml_legend=1 00:17:04.335 --rc geninfo_all_blocks=1 00:17:04.335 --rc geninfo_unexecuted_blocks=1 00:17:04.335 00:17:04.335 ' 00:17:04.335 15:21:32 spdkcli_raid -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:17:04.335 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:04.335 --rc genhtml_branch_coverage=1 00:17:04.335 --rc genhtml_function_coverage=1 00:17:04.335 --rc genhtml_legend=1 00:17:04.335 --rc geninfo_all_blocks=1 00:17:04.335 --rc geninfo_unexecuted_blocks=1 00:17:04.335 00:17:04.335 ' 00:17:04.335 15:21:32 spdkcli_raid -- spdkcli/raid.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:17:04.335 15:21:32 spdkcli_raid -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:17:04.335 15:21:32 spdkcli_raid -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:17:04.335 15:21:32 spdkcli_raid -- spdkcli/raid.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/iscsi_tgt/common.sh 00:17:04.335 15:21:32 spdkcli_raid -- iscsi_tgt/common.sh@9 -- # ISCSI_BRIDGE=iscsi_br 00:17:04.335 15:21:32 spdkcli_raid -- iscsi_tgt/common.sh@10 -- # INITIATOR_INTERFACE=spdk_init_int 00:17:04.335 15:21:32 spdkcli_raid -- iscsi_tgt/common.sh@11 -- # INITIATOR_BRIDGE=init_br 00:17:04.335 15:21:32 spdkcli_raid -- iscsi_tgt/common.sh@12 -- # TARGET_NAMESPACE=spdk_iscsi_ns 00:17:04.335 15:21:32 spdkcli_raid -- iscsi_tgt/common.sh@13 -- # TARGET_NS_CMD=(ip netns exec "$TARGET_NAMESPACE") 00:17:04.335 15:21:32 spdkcli_raid -- iscsi_tgt/common.sh@14 -- # TARGET_INTERFACE=spdk_tgt_int 00:17:04.335 15:21:32 spdkcli_raid -- iscsi_tgt/common.sh@15 -- # TARGET_INTERFACE2=spdk_tgt_int2 00:17:04.335 15:21:32 spdkcli_raid -- iscsi_tgt/common.sh@16 -- # TARGET_BRIDGE=tgt_br 00:17:04.335 15:21:32 spdkcli_raid -- iscsi_tgt/common.sh@17 -- # TARGET_BRIDGE2=tgt_br2 00:17:04.335 15:21:32 spdkcli_raid -- iscsi_tgt/common.sh@20 -- # TARGET_IP=10.0.0.1 00:17:04.335 15:21:32 spdkcli_raid -- iscsi_tgt/common.sh@21 -- # TARGET_IP2=10.0.0.3 00:17:04.335 15:21:32 spdkcli_raid -- iscsi_tgt/common.sh@22 -- # INITIATOR_IP=10.0.0.2 00:17:04.335 15:21:32 spdkcli_raid -- iscsi_tgt/common.sh@23 -- # ISCSI_PORT=3260 00:17:04.335 15:21:32 spdkcli_raid -- iscsi_tgt/common.sh@24 -- # NETMASK=10.0.0.2/32 00:17:04.335 15:21:32 spdkcli_raid -- iscsi_tgt/common.sh@25 -- # INITIATOR_TAG=2 00:17:04.335 15:21:32 spdkcli_raid -- iscsi_tgt/common.sh@26 -- # INITIATOR_NAME=ANY 00:17:04.335 15:21:32 spdkcli_raid -- iscsi_tgt/common.sh@27 -- # PORTAL_TAG=1 00:17:04.335 15:21:32 spdkcli_raid -- iscsi_tgt/common.sh@28 -- # ISCSI_APP=("${TARGET_NS_CMD[@]}" "${ISCSI_APP[@]}") 00:17:04.335 15:21:32 spdkcli_raid -- iscsi_tgt/common.sh@29 -- # ISCSI_TEST_CORE_MASK=0xF 00:17:04.335 15:21:32 spdkcli_raid -- spdkcli/raid.sh@12 -- # MATCH_FILE=spdkcli_raid.test 00:17:04.335 15:21:32 spdkcli_raid -- spdkcli/raid.sh@13 -- # SPDKCLI_BRANCH=/bdevs 00:17:04.335 15:21:32 spdkcli_raid -- spdkcli/raid.sh@14 -- # dirname /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:17:04.335 15:21:32 spdkcli_raid -- spdkcli/raid.sh@14 -- # readlink -f /home/vagrant/spdk_repo/spdk/test/spdkcli 00:17:04.335 15:21:32 spdkcli_raid -- spdkcli/raid.sh@14 -- # testdir=/home/vagrant/spdk_repo/spdk/test/spdkcli 00:17:04.335 15:21:32 spdkcli_raid -- spdkcli/raid.sh@15 -- # . /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:17:04.335 15:21:32 spdkcli_raid -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:17:04.335 15:21:32 spdkcli_raid -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:17:04.335 15:21:32 spdkcli_raid -- spdkcli/raid.sh@17 -- # trap cleanup EXIT 00:17:04.335 15:21:32 spdkcli_raid -- spdkcli/raid.sh@19 -- # timing_enter run_spdk_tgt 00:17:04.335 15:21:32 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:17:04.335 15:21:32 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:04.335 15:21:32 spdkcli_raid -- spdkcli/raid.sh@20 -- # run_spdk_tgt 00:17:04.335 15:21:32 spdkcli_raid -- spdkcli/common.sh@27 -- # spdk_tgt_pid=100207 00:17:04.335 15:21:32 spdkcli_raid -- spdkcli/common.sh@26 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x3 -p 0 00:17:04.335 15:21:32 spdkcli_raid -- spdkcli/common.sh@28 -- # waitforlisten 100207 00:17:04.335 15:21:32 spdkcli_raid -- common/autotest_common.sh@835 -- # '[' -z 100207 ']' 00:17:04.335 15:21:32 spdkcli_raid -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:04.335 15:21:32 spdkcli_raid -- common/autotest_common.sh@840 -- # local max_retries=100 00:17:04.335 15:21:32 spdkcli_raid -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:04.335 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:04.335 15:21:32 spdkcli_raid -- common/autotest_common.sh@844 -- # xtrace_disable 00:17:04.335 15:21:32 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:04.335 [2024-11-27 15:21:32.371840] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:17:04.335 [2024-11-27 15:21:32.372470] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100207 ] 00:17:04.595 [2024-11-27 15:21:32.540698] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:17:04.595 [2024-11-27 15:21:32.583506] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:04.595 [2024-11-27 15:21:32.583588] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:17:05.165 15:21:33 spdkcli_raid -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:17:05.165 15:21:33 spdkcli_raid -- common/autotest_common.sh@868 -- # return 0 00:17:05.165 15:21:33 spdkcli_raid -- spdkcli/raid.sh@21 -- # timing_exit run_spdk_tgt 00:17:05.165 15:21:33 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:17:05.165 15:21:33 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:05.165 15:21:33 spdkcli_raid -- spdkcli/raid.sh@23 -- # timing_enter spdkcli_create_malloc 00:17:05.165 15:21:33 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:17:05.165 15:21:33 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:05.165 15:21:33 spdkcli_raid -- spdkcli/raid.sh@26 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/malloc create 8 512 Malloc1'\'' '\''Malloc1'\'' True 00:17:05.165 '\''/bdevs/malloc create 8 512 Malloc2'\'' '\''Malloc2'\'' True 00:17:05.165 ' 00:17:07.072 Executing command: ['/bdevs/malloc create 8 512 Malloc1', 'Malloc1', True] 00:17:07.072 Executing command: ['/bdevs/malloc create 8 512 Malloc2', 'Malloc2', True] 00:17:07.072 15:21:34 spdkcli_raid -- spdkcli/raid.sh@27 -- # timing_exit spdkcli_create_malloc 00:17:07.072 15:21:34 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:17:07.072 15:21:34 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:07.072 15:21:34 spdkcli_raid -- spdkcli/raid.sh@29 -- # timing_enter spdkcli_create_raid 00:17:07.072 15:21:34 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:17:07.072 15:21:34 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:07.072 15:21:34 spdkcli_raid -- spdkcli/raid.sh@31 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/raid_volume create testraid 0 "Malloc1 Malloc2" 4'\'' '\''testraid'\'' True 00:17:07.072 ' 00:17:08.012 Executing command: ['/bdevs/raid_volume create testraid 0 "Malloc1 Malloc2" 4', 'testraid', True] 00:17:08.012 15:21:36 spdkcli_raid -- spdkcli/raid.sh@32 -- # timing_exit spdkcli_create_raid 00:17:08.012 15:21:36 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:17:08.012 15:21:36 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:08.012 15:21:36 spdkcli_raid -- spdkcli/raid.sh@34 -- # timing_enter spdkcli_check_match 00:17:08.012 15:21:36 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:17:08.012 15:21:36 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:08.012 15:21:36 spdkcli_raid -- spdkcli/raid.sh@35 -- # check_match 00:17:08.012 15:21:36 spdkcli_raid -- spdkcli/common.sh@44 -- # /home/vagrant/spdk_repo/spdk/scripts/spdkcli.py ll /bdevs 00:17:08.582 15:21:36 spdkcli_raid -- spdkcli/common.sh@45 -- # /home/vagrant/spdk_repo/spdk/test/app/match/match /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_raid.test.match 00:17:08.582 15:21:36 spdkcli_raid -- spdkcli/common.sh@46 -- # rm -f /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_raid.test 00:17:08.582 15:21:36 spdkcli_raid -- spdkcli/raid.sh@36 -- # timing_exit spdkcli_check_match 00:17:08.582 15:21:36 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:17:08.582 15:21:36 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:08.582 15:21:36 spdkcli_raid -- spdkcli/raid.sh@38 -- # timing_enter spdkcli_delete_raid 00:17:08.582 15:21:36 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:17:08.582 15:21:36 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:08.582 15:21:36 spdkcli_raid -- spdkcli/raid.sh@40 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/raid_volume delete testraid'\'' '\'''\'' True 00:17:08.582 ' 00:17:09.522 Executing command: ['/bdevs/raid_volume delete testraid', '', True] 00:17:09.782 15:21:37 spdkcli_raid -- spdkcli/raid.sh@41 -- # timing_exit spdkcli_delete_raid 00:17:09.782 15:21:37 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:17:09.782 15:21:37 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:09.782 15:21:37 spdkcli_raid -- spdkcli/raid.sh@43 -- # timing_enter spdkcli_delete_malloc 00:17:09.782 15:21:37 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:17:09.782 15:21:37 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:09.782 15:21:37 spdkcli_raid -- spdkcli/raid.sh@46 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/malloc delete Malloc1'\'' '\'''\'' True 00:17:09.782 '\''/bdevs/malloc delete Malloc2'\'' '\'''\'' True 00:17:09.782 ' 00:17:11.172 Executing command: ['/bdevs/malloc delete Malloc1', '', True] 00:17:11.172 Executing command: ['/bdevs/malloc delete Malloc2', '', True] 00:17:11.172 15:21:39 spdkcli_raid -- spdkcli/raid.sh@47 -- # timing_exit spdkcli_delete_malloc 00:17:11.172 15:21:39 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:17:11.172 15:21:39 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:11.172 15:21:39 spdkcli_raid -- spdkcli/raid.sh@49 -- # killprocess 100207 00:17:11.172 15:21:39 spdkcli_raid -- common/autotest_common.sh@954 -- # '[' -z 100207 ']' 00:17:11.172 15:21:39 spdkcli_raid -- common/autotest_common.sh@958 -- # kill -0 100207 00:17:11.172 15:21:39 spdkcli_raid -- common/autotest_common.sh@959 -- # uname 00:17:11.434 15:21:39 spdkcli_raid -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:17:11.434 15:21:39 spdkcli_raid -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 100207 00:17:11.434 15:21:39 spdkcli_raid -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:17:11.434 15:21:39 spdkcli_raid -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:17:11.434 15:21:39 spdkcli_raid -- common/autotest_common.sh@972 -- # echo 'killing process with pid 100207' 00:17:11.434 killing process with pid 100207 00:17:11.434 15:21:39 spdkcli_raid -- common/autotest_common.sh@973 -- # kill 100207 00:17:11.434 15:21:39 spdkcli_raid -- common/autotest_common.sh@978 -- # wait 100207 00:17:12.004 15:21:39 spdkcli_raid -- spdkcli/raid.sh@1 -- # cleanup 00:17:12.004 15:21:39 spdkcli_raid -- spdkcli/common.sh@10 -- # '[' -n 100207 ']' 00:17:12.004 15:21:39 spdkcli_raid -- spdkcli/common.sh@11 -- # killprocess 100207 00:17:12.004 15:21:39 spdkcli_raid -- common/autotest_common.sh@954 -- # '[' -z 100207 ']' 00:17:12.004 15:21:39 spdkcli_raid -- common/autotest_common.sh@958 -- # kill -0 100207 00:17:12.004 Process with pid 100207 is not found 00:17:12.004 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 958: kill: (100207) - No such process 00:17:12.004 15:21:39 spdkcli_raid -- common/autotest_common.sh@981 -- # echo 'Process with pid 100207 is not found' 00:17:12.004 15:21:39 spdkcli_raid -- spdkcli/common.sh@13 -- # '[' -n '' ']' 00:17:12.004 15:21:39 spdkcli_raid -- spdkcli/common.sh@16 -- # '[' -n '' ']' 00:17:12.004 15:21:39 spdkcli_raid -- spdkcli/common.sh@19 -- # '[' -n '' ']' 00:17:12.004 15:21:39 spdkcli_raid -- spdkcli/common.sh@22 -- # rm -f /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_raid.test /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_details_vhost.test /tmp/sample_aio 00:17:12.004 ************************************ 00:17:12.004 END TEST spdkcli_raid 00:17:12.004 ************************************ 00:17:12.004 00:17:12.004 real 0m7.919s 00:17:12.004 user 0m16.525s 00:17:12.004 sys 0m1.197s 00:17:12.004 15:21:39 spdkcli_raid -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:12.004 15:21:39 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:12.004 15:21:40 -- spdk/autotest.sh@191 -- # run_test blockdev_raid5f /home/vagrant/spdk_repo/spdk/test/bdev/blockdev.sh raid5f 00:17:12.004 15:21:40 -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:17:12.004 15:21:40 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:12.004 15:21:40 -- common/autotest_common.sh@10 -- # set +x 00:17:12.004 ************************************ 00:17:12.004 START TEST blockdev_raid5f 00:17:12.004 ************************************ 00:17:12.004 15:21:40 blockdev_raid5f -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/bdev/blockdev.sh raid5f 00:17:12.265 * Looking for test storage... 00:17:12.265 * Found test storage at /home/vagrant/spdk_repo/spdk/test/bdev 00:17:12.265 15:21:40 blockdev_raid5f -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:17:12.265 15:21:40 blockdev_raid5f -- common/autotest_common.sh@1693 -- # lcov --version 00:17:12.265 15:21:40 blockdev_raid5f -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:17:12.265 15:21:40 blockdev_raid5f -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:17:12.265 15:21:40 blockdev_raid5f -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:17:12.265 15:21:40 blockdev_raid5f -- scripts/common.sh@333 -- # local ver1 ver1_l 00:17:12.265 15:21:40 blockdev_raid5f -- scripts/common.sh@334 -- # local ver2 ver2_l 00:17:12.265 15:21:40 blockdev_raid5f -- scripts/common.sh@336 -- # IFS=.-: 00:17:12.265 15:21:40 blockdev_raid5f -- scripts/common.sh@336 -- # read -ra ver1 00:17:12.265 15:21:40 blockdev_raid5f -- scripts/common.sh@337 -- # IFS=.-: 00:17:12.265 15:21:40 blockdev_raid5f -- scripts/common.sh@337 -- # read -ra ver2 00:17:12.265 15:21:40 blockdev_raid5f -- scripts/common.sh@338 -- # local 'op=<' 00:17:12.265 15:21:40 blockdev_raid5f -- scripts/common.sh@340 -- # ver1_l=2 00:17:12.265 15:21:40 blockdev_raid5f -- scripts/common.sh@341 -- # ver2_l=1 00:17:12.265 15:21:40 blockdev_raid5f -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:17:12.265 15:21:40 blockdev_raid5f -- scripts/common.sh@344 -- # case "$op" in 00:17:12.265 15:21:40 blockdev_raid5f -- scripts/common.sh@345 -- # : 1 00:17:12.265 15:21:40 blockdev_raid5f -- scripts/common.sh@364 -- # (( v = 0 )) 00:17:12.265 15:21:40 blockdev_raid5f -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:17:12.265 15:21:40 blockdev_raid5f -- scripts/common.sh@365 -- # decimal 1 00:17:12.265 15:21:40 blockdev_raid5f -- scripts/common.sh@353 -- # local d=1 00:17:12.265 15:21:40 blockdev_raid5f -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:17:12.265 15:21:40 blockdev_raid5f -- scripts/common.sh@355 -- # echo 1 00:17:12.265 15:21:40 blockdev_raid5f -- scripts/common.sh@365 -- # ver1[v]=1 00:17:12.265 15:21:40 blockdev_raid5f -- scripts/common.sh@366 -- # decimal 2 00:17:12.265 15:21:40 blockdev_raid5f -- scripts/common.sh@353 -- # local d=2 00:17:12.265 15:21:40 blockdev_raid5f -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:17:12.265 15:21:40 blockdev_raid5f -- scripts/common.sh@355 -- # echo 2 00:17:12.265 15:21:40 blockdev_raid5f -- scripts/common.sh@366 -- # ver2[v]=2 00:17:12.265 15:21:40 blockdev_raid5f -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:17:12.265 15:21:40 blockdev_raid5f -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:17:12.265 15:21:40 blockdev_raid5f -- scripts/common.sh@368 -- # return 0 00:17:12.265 15:21:40 blockdev_raid5f -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:17:12.265 15:21:40 blockdev_raid5f -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:17:12.265 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:12.265 --rc genhtml_branch_coverage=1 00:17:12.265 --rc genhtml_function_coverage=1 00:17:12.265 --rc genhtml_legend=1 00:17:12.265 --rc geninfo_all_blocks=1 00:17:12.265 --rc geninfo_unexecuted_blocks=1 00:17:12.265 00:17:12.265 ' 00:17:12.265 15:21:40 blockdev_raid5f -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:17:12.265 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:12.265 --rc genhtml_branch_coverage=1 00:17:12.265 --rc genhtml_function_coverage=1 00:17:12.265 --rc genhtml_legend=1 00:17:12.265 --rc geninfo_all_blocks=1 00:17:12.265 --rc geninfo_unexecuted_blocks=1 00:17:12.265 00:17:12.265 ' 00:17:12.265 15:21:40 blockdev_raid5f -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:17:12.265 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:12.265 --rc genhtml_branch_coverage=1 00:17:12.265 --rc genhtml_function_coverage=1 00:17:12.265 --rc genhtml_legend=1 00:17:12.265 --rc geninfo_all_blocks=1 00:17:12.265 --rc geninfo_unexecuted_blocks=1 00:17:12.265 00:17:12.265 ' 00:17:12.265 15:21:40 blockdev_raid5f -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:17:12.265 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:12.265 --rc genhtml_branch_coverage=1 00:17:12.265 --rc genhtml_function_coverage=1 00:17:12.265 --rc genhtml_legend=1 00:17:12.265 --rc geninfo_all_blocks=1 00:17:12.265 --rc geninfo_unexecuted_blocks=1 00:17:12.265 00:17:12.265 ' 00:17:12.265 15:21:40 blockdev_raid5f -- bdev/blockdev.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:17:12.265 15:21:40 blockdev_raid5f -- bdev/nbd_common.sh@6 -- # set -e 00:17:12.265 15:21:40 blockdev_raid5f -- bdev/blockdev.sh@12 -- # rpc_py=rpc_cmd 00:17:12.265 15:21:40 blockdev_raid5f -- bdev/blockdev.sh@13 -- # conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:17:12.266 15:21:40 blockdev_raid5f -- bdev/blockdev.sh@14 -- # nonenclosed_conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json 00:17:12.266 15:21:40 blockdev_raid5f -- bdev/blockdev.sh@15 -- # nonarray_conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json 00:17:12.266 15:21:40 blockdev_raid5f -- bdev/blockdev.sh@17 -- # export RPC_PIPE_TIMEOUT=30 00:17:12.266 15:21:40 blockdev_raid5f -- bdev/blockdev.sh@17 -- # RPC_PIPE_TIMEOUT=30 00:17:12.266 15:21:40 blockdev_raid5f -- bdev/blockdev.sh@20 -- # : 00:17:12.266 15:21:40 blockdev_raid5f -- bdev/blockdev.sh@707 -- # QOS_DEV_1=Malloc_0 00:17:12.266 15:21:40 blockdev_raid5f -- bdev/blockdev.sh@708 -- # QOS_DEV_2=Null_1 00:17:12.266 15:21:40 blockdev_raid5f -- bdev/blockdev.sh@709 -- # QOS_RUN_TIME=5 00:17:12.266 15:21:40 blockdev_raid5f -- bdev/blockdev.sh@711 -- # uname -s 00:17:12.266 15:21:40 blockdev_raid5f -- bdev/blockdev.sh@711 -- # '[' Linux = Linux ']' 00:17:12.266 15:21:40 blockdev_raid5f -- bdev/blockdev.sh@713 -- # PRE_RESERVED_MEM=0 00:17:12.266 15:21:40 blockdev_raid5f -- bdev/blockdev.sh@719 -- # test_type=raid5f 00:17:12.266 15:21:40 blockdev_raid5f -- bdev/blockdev.sh@720 -- # crypto_device= 00:17:12.266 15:21:40 blockdev_raid5f -- bdev/blockdev.sh@721 -- # dek= 00:17:12.266 15:21:40 blockdev_raid5f -- bdev/blockdev.sh@722 -- # env_ctx= 00:17:12.266 15:21:40 blockdev_raid5f -- bdev/blockdev.sh@723 -- # wait_for_rpc= 00:17:12.266 15:21:40 blockdev_raid5f -- bdev/blockdev.sh@724 -- # '[' -n '' ']' 00:17:12.266 15:21:40 blockdev_raid5f -- bdev/blockdev.sh@727 -- # [[ raid5f == bdev ]] 00:17:12.266 15:21:40 blockdev_raid5f -- bdev/blockdev.sh@727 -- # [[ raid5f == crypto_* ]] 00:17:12.266 15:21:40 blockdev_raid5f -- bdev/blockdev.sh@730 -- # start_spdk_tgt 00:17:12.266 15:21:40 blockdev_raid5f -- bdev/blockdev.sh@47 -- # spdk_tgt_pid=100465 00:17:12.266 15:21:40 blockdev_raid5f -- bdev/blockdev.sh@46 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt '' '' 00:17:12.266 15:21:40 blockdev_raid5f -- bdev/blockdev.sh@48 -- # trap 'killprocess "$spdk_tgt_pid"; exit 1' SIGINT SIGTERM EXIT 00:17:12.266 15:21:40 blockdev_raid5f -- bdev/blockdev.sh@49 -- # waitforlisten 100465 00:17:12.266 15:21:40 blockdev_raid5f -- common/autotest_common.sh@835 -- # '[' -z 100465 ']' 00:17:12.266 15:21:40 blockdev_raid5f -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:12.266 15:21:40 blockdev_raid5f -- common/autotest_common.sh@840 -- # local max_retries=100 00:17:12.266 15:21:40 blockdev_raid5f -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:12.266 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:12.266 15:21:40 blockdev_raid5f -- common/autotest_common.sh@844 -- # xtrace_disable 00:17:12.266 15:21:40 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:12.266 [2024-11-27 15:21:40.363901] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:17:12.266 [2024-11-27 15:21:40.364141] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100465 ] 00:17:12.526 [2024-11-27 15:21:40.547317] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:12.526 [2024-11-27 15:21:40.588620] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:13.097 15:21:41 blockdev_raid5f -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:17:13.097 15:21:41 blockdev_raid5f -- common/autotest_common.sh@868 -- # return 0 00:17:13.097 15:21:41 blockdev_raid5f -- bdev/blockdev.sh@731 -- # case "$test_type" in 00:17:13.097 15:21:41 blockdev_raid5f -- bdev/blockdev.sh@763 -- # setup_raid5f_conf 00:17:13.097 15:21:41 blockdev_raid5f -- bdev/blockdev.sh@279 -- # rpc_cmd 00:17:13.097 15:21:41 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:13.097 15:21:41 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:13.097 Malloc0 00:17:13.357 Malloc1 00:17:13.357 Malloc2 00:17:13.357 15:21:41 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:13.357 15:21:41 blockdev_raid5f -- bdev/blockdev.sh@774 -- # rpc_cmd bdev_wait_for_examine 00:17:13.357 15:21:41 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:13.357 15:21:41 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:13.357 15:21:41 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:13.357 15:21:41 blockdev_raid5f -- bdev/blockdev.sh@777 -- # cat 00:17:13.357 15:21:41 blockdev_raid5f -- bdev/blockdev.sh@777 -- # rpc_cmd save_subsystem_config -n accel 00:17:13.357 15:21:41 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:13.357 15:21:41 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:13.357 15:21:41 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:13.357 15:21:41 blockdev_raid5f -- bdev/blockdev.sh@777 -- # rpc_cmd save_subsystem_config -n bdev 00:17:13.357 15:21:41 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:13.357 15:21:41 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:13.357 15:21:41 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:13.357 15:21:41 blockdev_raid5f -- bdev/blockdev.sh@777 -- # rpc_cmd save_subsystem_config -n iobuf 00:17:13.357 15:21:41 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:13.357 15:21:41 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:13.357 15:21:41 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:13.357 15:21:41 blockdev_raid5f -- bdev/blockdev.sh@785 -- # mapfile -t bdevs 00:17:13.357 15:21:41 blockdev_raid5f -- bdev/blockdev.sh@785 -- # rpc_cmd bdev_get_bdevs 00:17:13.357 15:21:41 blockdev_raid5f -- bdev/blockdev.sh@785 -- # jq -r '.[] | select(.claimed == false)' 00:17:13.357 15:21:41 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:13.357 15:21:41 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:13.357 15:21:41 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:13.357 15:21:41 blockdev_raid5f -- bdev/blockdev.sh@786 -- # mapfile -t bdevs_name 00:17:13.357 15:21:41 blockdev_raid5f -- bdev/blockdev.sh@786 -- # jq -r .name 00:17:13.357 15:21:41 blockdev_raid5f -- bdev/blockdev.sh@786 -- # printf '%s\n' '{' ' "name": "raid5f",' ' "aliases": [' ' "f18e5124-227c-4909-b5ec-1ecded510ce2"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 131072,' ' "uuid": "f18e5124-227c-4909-b5ec-1ecded510ce2",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": false,' ' "flush": false,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "raid": {' ' "uuid": "f18e5124-227c-4909-b5ec-1ecded510ce2",' ' "strip_size_kb": 2,' ' "state": "online",' ' "raid_level": "raid5f",' ' "superblock": false,' ' "num_base_bdevs": 3,' ' "num_base_bdevs_discovered": 3,' ' "num_base_bdevs_operational": 3,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc0",' ' "uuid": "514fae21-1a69-422a-be2a-da58c5d8630a",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc1",' ' "uuid": "0bfe0434-02fc-4d29-a853-8a8573454675",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc2",' ' "uuid": "eb56b52a-0b23-456a-b81a-8f03b15a6b50",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' 00:17:13.357 15:21:41 blockdev_raid5f -- bdev/blockdev.sh@787 -- # bdev_list=("${bdevs_name[@]}") 00:17:13.357 15:21:41 blockdev_raid5f -- bdev/blockdev.sh@789 -- # hello_world_bdev=raid5f 00:17:13.357 15:21:41 blockdev_raid5f -- bdev/blockdev.sh@790 -- # trap - SIGINT SIGTERM EXIT 00:17:13.357 15:21:41 blockdev_raid5f -- bdev/blockdev.sh@791 -- # killprocess 100465 00:17:13.357 15:21:41 blockdev_raid5f -- common/autotest_common.sh@954 -- # '[' -z 100465 ']' 00:17:13.357 15:21:41 blockdev_raid5f -- common/autotest_common.sh@958 -- # kill -0 100465 00:17:13.357 15:21:41 blockdev_raid5f -- common/autotest_common.sh@959 -- # uname 00:17:13.357 15:21:41 blockdev_raid5f -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:17:13.357 15:21:41 blockdev_raid5f -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 100465 00:17:13.618 killing process with pid 100465 00:17:13.618 15:21:41 blockdev_raid5f -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:17:13.618 15:21:41 blockdev_raid5f -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:17:13.618 15:21:41 blockdev_raid5f -- common/autotest_common.sh@972 -- # echo 'killing process with pid 100465' 00:17:13.618 15:21:41 blockdev_raid5f -- common/autotest_common.sh@973 -- # kill 100465 00:17:13.618 15:21:41 blockdev_raid5f -- common/autotest_common.sh@978 -- # wait 100465 00:17:14.189 15:21:42 blockdev_raid5f -- bdev/blockdev.sh@795 -- # trap cleanup SIGINT SIGTERM EXIT 00:17:14.189 15:21:42 blockdev_raid5f -- bdev/blockdev.sh@797 -- # run_test bdev_hello_world /home/vagrant/spdk_repo/spdk/build/examples/hello_bdev --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -b raid5f '' 00:17:14.189 15:21:42 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:17:14.189 15:21:42 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:14.189 15:21:42 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:14.189 ************************************ 00:17:14.189 START TEST bdev_hello_world 00:17:14.189 ************************************ 00:17:14.189 15:21:42 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/hello_bdev --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -b raid5f '' 00:17:14.189 [2024-11-27 15:21:42.226682] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:17:14.189 [2024-11-27 15:21:42.226803] [ DPDK EAL parameters: hello_bdev --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100510 ] 00:17:14.448 [2024-11-27 15:21:42.395487] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:14.448 [2024-11-27 15:21:42.436354] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:14.707 [2024-11-27 15:21:42.682875] hello_bdev.c: 222:hello_start: *NOTICE*: Successfully started the application 00:17:14.707 [2024-11-27 15:21:42.682936] hello_bdev.c: 231:hello_start: *NOTICE*: Opening the bdev raid5f 00:17:14.707 [2024-11-27 15:21:42.682961] hello_bdev.c: 244:hello_start: *NOTICE*: Opening io channel 00:17:14.707 [2024-11-27 15:21:42.683275] hello_bdev.c: 138:hello_write: *NOTICE*: Writing to the bdev 00:17:14.707 [2024-11-27 15:21:42.683412] hello_bdev.c: 117:write_complete: *NOTICE*: bdev io write completed successfully 00:17:14.707 [2024-11-27 15:21:42.683427] hello_bdev.c: 84:hello_read: *NOTICE*: Reading io 00:17:14.707 [2024-11-27 15:21:42.683483] hello_bdev.c: 65:read_complete: *NOTICE*: Read string from bdev : Hello World! 00:17:14.707 00:17:14.707 [2024-11-27 15:21:42.683502] hello_bdev.c: 74:read_complete: *NOTICE*: Stopping app 00:17:14.968 00:17:14.968 real 0m0.903s 00:17:14.968 user 0m0.515s 00:17:14.968 sys 0m0.281s 00:17:14.968 15:21:43 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:14.968 15:21:43 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@10 -- # set +x 00:17:14.968 ************************************ 00:17:14.968 END TEST bdev_hello_world 00:17:14.968 ************************************ 00:17:15.233 15:21:43 blockdev_raid5f -- bdev/blockdev.sh@798 -- # run_test bdev_bounds bdev_bounds '' 00:17:15.233 15:21:43 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:17:15.233 15:21:43 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:15.233 15:21:43 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:15.233 ************************************ 00:17:15.233 START TEST bdev_bounds 00:17:15.233 ************************************ 00:17:15.233 15:21:43 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@1129 -- # bdev_bounds '' 00:17:15.233 15:21:43 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@289 -- # bdevio_pid=100541 00:17:15.233 15:21:43 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@288 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdevio/bdevio -w -s 0 --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json '' 00:17:15.233 15:21:43 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@290 -- # trap 'cleanup; killprocess $bdevio_pid; exit 1' SIGINT SIGTERM EXIT 00:17:15.233 Process bdevio pid: 100541 00:17:15.233 15:21:43 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@291 -- # echo 'Process bdevio pid: 100541' 00:17:15.233 15:21:43 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@292 -- # waitforlisten 100541 00:17:15.233 15:21:43 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@835 -- # '[' -z 100541 ']' 00:17:15.233 15:21:43 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:15.233 15:21:43 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@840 -- # local max_retries=100 00:17:15.233 15:21:43 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:15.233 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:15.233 15:21:43 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@844 -- # xtrace_disable 00:17:15.233 15:21:43 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@10 -- # set +x 00:17:15.233 [2024-11-27 15:21:43.223283] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:17:15.233 [2024-11-27 15:21:43.223535] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 -m 0 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100541 ] 00:17:15.493 [2024-11-27 15:21:43.398377] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:17:15.493 [2024-11-27 15:21:43.442781] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:17:15.493 [2024-11-27 15:21:43.442895] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:15.493 [2024-11-27 15:21:43.443002] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:17:16.064 15:21:44 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:17:16.064 15:21:44 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@868 -- # return 0 00:17:16.064 15:21:44 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@293 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdevio/tests.py perform_tests 00:17:16.064 I/O targets: 00:17:16.064 raid5f: 131072 blocks of 512 bytes (64 MiB) 00:17:16.064 00:17:16.064 00:17:16.064 CUnit - A unit testing framework for C - Version 2.1-3 00:17:16.064 http://cunit.sourceforge.net/ 00:17:16.064 00:17:16.064 00:17:16.064 Suite: bdevio tests on: raid5f 00:17:16.064 Test: blockdev write read block ...passed 00:17:16.064 Test: blockdev write zeroes read block ...passed 00:17:16.064 Test: blockdev write zeroes read no split ...passed 00:17:16.325 Test: blockdev write zeroes read split ...passed 00:17:16.325 Test: blockdev write zeroes read split partial ...passed 00:17:16.325 Test: blockdev reset ...passed 00:17:16.325 Test: blockdev write read 8 blocks ...passed 00:17:16.325 Test: blockdev write read size > 128k ...passed 00:17:16.325 Test: blockdev write read invalid size ...passed 00:17:16.325 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:17:16.325 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:17:16.325 Test: blockdev write read max offset ...passed 00:17:16.325 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:17:16.325 Test: blockdev writev readv 8 blocks ...passed 00:17:16.325 Test: blockdev writev readv 30 x 1block ...passed 00:17:16.325 Test: blockdev writev readv block ...passed 00:17:16.325 Test: blockdev writev readv size > 128k ...passed 00:17:16.325 Test: blockdev writev readv size > 128k in two iovs ...passed 00:17:16.325 Test: blockdev comparev and writev ...passed 00:17:16.325 Test: blockdev nvme passthru rw ...passed 00:17:16.325 Test: blockdev nvme passthru vendor specific ...passed 00:17:16.325 Test: blockdev nvme admin passthru ...passed 00:17:16.325 Test: blockdev copy ...passed 00:17:16.325 00:17:16.325 Run Summary: Type Total Ran Passed Failed Inactive 00:17:16.325 suites 1 1 n/a 0 0 00:17:16.325 tests 23 23 23 0 0 00:17:16.325 asserts 130 130 130 0 n/a 00:17:16.325 00:17:16.325 Elapsed time = 0.343 seconds 00:17:16.325 0 00:17:16.325 15:21:44 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@294 -- # killprocess 100541 00:17:16.325 15:21:44 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@954 -- # '[' -z 100541 ']' 00:17:16.325 15:21:44 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@958 -- # kill -0 100541 00:17:16.325 15:21:44 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@959 -- # uname 00:17:16.325 15:21:44 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:17:16.325 15:21:44 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 100541 00:17:16.325 15:21:44 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:17:16.325 15:21:44 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:17:16.325 15:21:44 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@972 -- # echo 'killing process with pid 100541' 00:17:16.325 killing process with pid 100541 00:17:16.325 15:21:44 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@973 -- # kill 100541 00:17:16.325 15:21:44 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@978 -- # wait 100541 00:17:16.896 15:21:44 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@295 -- # trap - SIGINT SIGTERM EXIT 00:17:16.896 00:17:16.896 real 0m1.600s 00:17:16.896 user 0m3.750s 00:17:16.896 sys 0m0.434s 00:17:16.896 15:21:44 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:16.896 15:21:44 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@10 -- # set +x 00:17:16.896 ************************************ 00:17:16.896 END TEST bdev_bounds 00:17:16.896 ************************************ 00:17:16.896 15:21:44 blockdev_raid5f -- bdev/blockdev.sh@799 -- # run_test bdev_nbd nbd_function_test /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json raid5f '' 00:17:16.896 15:21:44 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:17:16.896 15:21:44 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:16.896 15:21:44 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:16.896 ************************************ 00:17:16.896 START TEST bdev_nbd 00:17:16.896 ************************************ 00:17:16.896 15:21:44 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@1129 -- # nbd_function_test /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json raid5f '' 00:17:16.896 15:21:44 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@299 -- # uname -s 00:17:16.896 15:21:44 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@299 -- # [[ Linux == Linux ]] 00:17:16.896 15:21:44 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@301 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:16.896 15:21:44 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@302 -- # local conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:17:16.896 15:21:44 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@303 -- # bdev_all=('raid5f') 00:17:16.896 15:21:44 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@303 -- # local bdev_all 00:17:16.896 15:21:44 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@304 -- # local bdev_num=1 00:17:16.896 15:21:44 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@308 -- # [[ -e /sys/module/nbd ]] 00:17:16.897 15:21:44 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@310 -- # nbd_all=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11' '/dev/nbd12' '/dev/nbd13' '/dev/nbd14' '/dev/nbd15' '/dev/nbd2' '/dev/nbd3' '/dev/nbd4' '/dev/nbd5' '/dev/nbd6' '/dev/nbd7' '/dev/nbd8' '/dev/nbd9') 00:17:16.897 15:21:44 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@310 -- # local nbd_all 00:17:16.897 15:21:44 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@311 -- # bdev_num=1 00:17:16.897 15:21:44 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@313 -- # nbd_list=('/dev/nbd0') 00:17:16.897 15:21:44 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@313 -- # local nbd_list 00:17:16.897 15:21:44 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@314 -- # bdev_list=('raid5f') 00:17:16.897 15:21:44 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@314 -- # local bdev_list 00:17:16.897 15:21:44 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@317 -- # nbd_pid=100584 00:17:16.897 15:21:44 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@316 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-nbd.sock -i 0 --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json '' 00:17:16.897 15:21:44 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@318 -- # trap 'cleanup; killprocess $nbd_pid' SIGINT SIGTERM EXIT 00:17:16.897 15:21:44 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@319 -- # waitforlisten 100584 /var/tmp/spdk-nbd.sock 00:17:16.897 15:21:44 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@835 -- # '[' -z 100584 ']' 00:17:16.897 15:21:44 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:17:16.897 15:21:44 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@840 -- # local max_retries=100 00:17:16.897 15:21:44 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:17:16.897 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:17:16.897 15:21:44 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@844 -- # xtrace_disable 00:17:16.897 15:21:44 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@10 -- # set +x 00:17:16.897 [2024-11-27 15:21:44.902536] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:17:16.897 [2024-11-27 15:21:44.902734] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:17:17.157 [2024-11-27 15:21:45.075224] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:17.157 [2024-11-27 15:21:45.113523] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:17.729 15:21:45 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:17:17.729 15:21:45 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@868 -- # return 0 00:17:17.729 15:21:45 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@321 -- # nbd_rpc_start_stop_verify /var/tmp/spdk-nbd.sock raid5f 00:17:17.729 15:21:45 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@113 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:17.729 15:21:45 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@114 -- # bdev_list=('raid5f') 00:17:17.729 15:21:45 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@114 -- # local bdev_list 00:17:17.729 15:21:45 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@116 -- # nbd_start_disks_without_nbd_idx /var/tmp/spdk-nbd.sock raid5f 00:17:17.729 15:21:45 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@22 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:17.729 15:21:45 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@23 -- # bdev_list=('raid5f') 00:17:17.729 15:21:45 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@23 -- # local bdev_list 00:17:17.729 15:21:45 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@24 -- # local i 00:17:17.729 15:21:45 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@25 -- # local nbd_device 00:17:17.729 15:21:45 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i = 0 )) 00:17:17.729 15:21:45 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 1 )) 00:17:17.729 15:21:45 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk raid5f 00:17:17.989 15:21:45 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd0 00:17:17.989 15:21:45 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd0 00:17:17.989 15:21:45 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd0 00:17:17.989 15:21:45 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:17:17.989 15:21:45 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@873 -- # local i 00:17:17.989 15:21:45 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:17:17.989 15:21:45 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:17:17.989 15:21:45 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:17:17.989 15:21:45 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@877 -- # break 00:17:17.989 15:21:45 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:17:17.989 15:21:45 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:17:17.989 15:21:45 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:17:17.989 1+0 records in 00:17:17.989 1+0 records out 00:17:17.989 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00045897 s, 8.9 MB/s 00:17:17.989 15:21:45 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:17.989 15:21:45 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@890 -- # size=4096 00:17:17.989 15:21:45 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:17.989 15:21:45 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:17:17.989 15:21:45 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@893 -- # return 0 00:17:17.990 15:21:45 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:17:17.990 15:21:45 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 1 )) 00:17:17.990 15:21:45 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@118 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:17:18.250 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@118 -- # nbd_disks_json='[ 00:17:18.250 { 00:17:18.250 "nbd_device": "/dev/nbd0", 00:17:18.250 "bdev_name": "raid5f" 00:17:18.250 } 00:17:18.250 ]' 00:17:18.250 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # nbd_disks_name=($(echo "${nbd_disks_json}" | jq -r '.[] | .nbd_device')) 00:17:18.250 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # echo '[ 00:17:18.250 { 00:17:18.250 "nbd_device": "/dev/nbd0", 00:17:18.250 "bdev_name": "raid5f" 00:17:18.250 } 00:17:18.250 ]' 00:17:18.250 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # jq -r '.[] | .nbd_device' 00:17:18.250 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@120 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:17:18.250 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:18.250 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:17:18.250 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:17:18.250 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:17:18.250 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:17:18.250 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:17:18.511 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:17:18.511 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:17:18.511 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:17:18.511 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:17:18.511 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:17:18.511 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:17:18.511 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:17:18.511 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:17:18.511 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@122 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:17:18.511 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:18.511 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:17:18.774 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:17:18.774 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[]' 00:17:18.774 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:17:18.774 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:17:18.774 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '' 00:17:18.774 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:17:18.774 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # true 00:17:18.774 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=0 00:17:18.774 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 0 00:17:18.774 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@122 -- # count=0 00:17:18.774 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@123 -- # '[' 0 -ne 0 ']' 00:17:18.774 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@127 -- # return 0 00:17:18.774 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@322 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock raid5f /dev/nbd0 00:17:18.774 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:18.774 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@91 -- # bdev_list=('raid5f') 00:17:18.774 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@91 -- # local bdev_list 00:17:18.774 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0') 00:17:18.774 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@92 -- # local nbd_list 00:17:18.774 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock raid5f /dev/nbd0 00:17:18.774 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:18.774 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@10 -- # bdev_list=('raid5f') 00:17:18.774 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@10 -- # local bdev_list 00:17:18.774 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:17:18.774 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@11 -- # local nbd_list 00:17:18.774 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@12 -- # local i 00:17:18.775 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:17:18.775 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:17:18.775 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk raid5f /dev/nbd0 00:17:19.035 /dev/nbd0 00:17:19.035 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:17:19.035 15:21:46 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:17:19.035 15:21:46 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:17:19.035 15:21:46 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@873 -- # local i 00:17:19.035 15:21:46 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:17:19.035 15:21:46 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:17:19.035 15:21:46 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:17:19.035 15:21:46 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@877 -- # break 00:17:19.035 15:21:46 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:17:19.035 15:21:46 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:17:19.035 15:21:46 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:17:19.035 1+0 records in 00:17:19.035 1+0 records out 00:17:19.035 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000626939 s, 6.5 MB/s 00:17:19.035 15:21:46 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:19.035 15:21:46 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@890 -- # size=4096 00:17:19.035 15:21:46 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:19.035 15:21:47 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:17:19.035 15:21:47 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@893 -- # return 0 00:17:19.035 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:17:19.035 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:17:19.035 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:17:19.035 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:19.035 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:17:19.295 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:17:19.295 { 00:17:19.295 "nbd_device": "/dev/nbd0", 00:17:19.295 "bdev_name": "raid5f" 00:17:19.295 } 00:17:19.295 ]' 00:17:19.295 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:17:19.295 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[ 00:17:19.295 { 00:17:19.295 "nbd_device": "/dev/nbd0", 00:17:19.295 "bdev_name": "raid5f" 00:17:19.295 } 00:17:19.295 ]' 00:17:19.295 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:17:19.295 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:17:19.295 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:17:19.295 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=1 00:17:19.295 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 1 00:17:19.295 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@95 -- # count=1 00:17:19.295 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@96 -- # '[' 1 -ne 1 ']' 00:17:19.295 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify /dev/nbd0 write 00:17:19.295 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0') 00:17:19.295 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # local nbd_list 00:17:19.295 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@71 -- # local operation=write 00:17:19.296 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:17:19.296 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:17:19.296 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest bs=4096 count=256 00:17:19.296 256+0 records in 00:17:19.296 256+0 records out 00:17:19.296 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0130916 s, 80.1 MB/s 00:17:19.296 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:17:19.296 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:17:19.296 256+0 records in 00:17:19.296 256+0 records out 00:17:19.296 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0303687 s, 34.5 MB/s 00:17:19.296 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify /dev/nbd0 verify 00:17:19.296 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0') 00:17:19.296 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # local nbd_list 00:17:19.296 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@71 -- # local operation=verify 00:17:19.296 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:17:19.296 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:17:19.296 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:17:19.296 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:17:19.296 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest /dev/nbd0 00:17:19.296 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:17:19.296 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:17:19.296 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:19.296 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:17:19.296 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:17:19.296 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:17:19.296 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:17:19.296 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:17:19.555 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:17:19.555 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:17:19.555 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:17:19.555 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:17:19.555 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:17:19.555 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:17:19.555 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:17:19.555 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:17:19.555 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:17:19.556 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:19.556 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:17:19.815 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:17:19.815 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:17:19.815 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[]' 00:17:19.815 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:17:19.815 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '' 00:17:19.815 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:17:19.815 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # true 00:17:19.815 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=0 00:17:19.815 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 0 00:17:19.815 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@104 -- # count=0 00:17:19.815 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:17:19.815 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@109 -- # return 0 00:17:19.815 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@323 -- # nbd_with_lvol_verify /var/tmp/spdk-nbd.sock /dev/nbd0 00:17:19.815 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@131 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:19.815 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@132 -- # local nbd=/dev/nbd0 00:17:19.815 15:21:47 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@134 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create -b malloc_lvol_verify 16 512 00:17:20.075 malloc_lvol_verify 00:17:20.075 15:21:48 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@135 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_lvol_create_lvstore malloc_lvol_verify lvs 00:17:20.335 cf1205c5-d97a-422e-934d-914451ee14c5 00:17:20.335 15:21:48 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@136 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_lvol_create lvol 4 -l lvs 00:17:20.335 28d71324-b32c-46fa-a702-a547a547fbf6 00:17:20.335 15:21:48 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@137 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk lvs/lvol /dev/nbd0 00:17:20.595 /dev/nbd0 00:17:20.595 15:21:48 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@139 -- # wait_for_nbd_set_capacity /dev/nbd0 00:17:20.595 15:21:48 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@146 -- # local nbd=nbd0 00:17:20.595 15:21:48 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@148 -- # [[ -e /sys/block/nbd0/size ]] 00:17:20.595 15:21:48 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@150 -- # (( 8192 == 0 )) 00:17:20.595 15:21:48 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@141 -- # mkfs.ext4 /dev/nbd0 00:17:20.595 mke2fs 1.47.0 (5-Feb-2023) 00:17:20.595 Discarding device blocks: 0/4096 done 00:17:20.595 Creating filesystem with 4096 1k blocks and 1024 inodes 00:17:20.595 00:17:20.595 Allocating group tables: 0/1 done 00:17:20.595 Writing inode tables: 0/1 done 00:17:20.595 Creating journal (1024 blocks): done 00:17:20.595 Writing superblocks and filesystem accounting information: 0/1 done 00:17:20.595 00:17:20.595 15:21:48 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@142 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:17:20.595 15:21:48 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:20.595 15:21:48 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:17:20.595 15:21:48 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:17:20.595 15:21:48 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:17:20.595 15:21:48 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:17:20.595 15:21:48 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:17:20.855 15:21:48 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:17:20.855 15:21:48 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:17:20.855 15:21:48 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:17:20.855 15:21:48 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:17:20.855 15:21:48 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:17:20.855 15:21:48 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:17:20.855 15:21:48 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:17:20.855 15:21:48 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:17:20.855 15:21:48 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@325 -- # killprocess 100584 00:17:20.855 15:21:48 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@954 -- # '[' -z 100584 ']' 00:17:20.855 15:21:48 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@958 -- # kill -0 100584 00:17:20.855 15:21:48 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@959 -- # uname 00:17:20.855 15:21:48 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:17:20.855 15:21:48 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 100584 00:17:20.855 15:21:48 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:17:20.855 killing process with pid 100584 00:17:20.855 15:21:48 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:17:20.855 15:21:48 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@972 -- # echo 'killing process with pid 100584' 00:17:20.855 15:21:48 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@973 -- # kill 100584 00:17:20.855 15:21:48 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@978 -- # wait 100584 00:17:21.426 15:21:49 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@326 -- # trap - SIGINT SIGTERM EXIT 00:17:21.426 00:17:21.426 real 0m4.485s 00:17:21.426 user 0m6.339s 00:17:21.426 sys 0m1.383s 00:17:21.426 15:21:49 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:21.426 15:21:49 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@10 -- # set +x 00:17:21.426 ************************************ 00:17:21.426 END TEST bdev_nbd 00:17:21.426 ************************************ 00:17:21.426 15:21:49 blockdev_raid5f -- bdev/blockdev.sh@800 -- # [[ y == y ]] 00:17:21.426 15:21:49 blockdev_raid5f -- bdev/blockdev.sh@801 -- # '[' raid5f = nvme ']' 00:17:21.426 15:21:49 blockdev_raid5f -- bdev/blockdev.sh@801 -- # '[' raid5f = gpt ']' 00:17:21.426 15:21:49 blockdev_raid5f -- bdev/blockdev.sh@805 -- # run_test bdev_fio fio_test_suite '' 00:17:21.426 15:21:49 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:17:21.426 15:21:49 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:21.426 15:21:49 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:21.426 ************************************ 00:17:21.426 START TEST bdev_fio 00:17:21.426 ************************************ 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1129 -- # fio_test_suite '' 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@330 -- # local env_context 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@334 -- # pushd /home/vagrant/spdk_repo/spdk/test/bdev 00:17:21.426 /home/vagrant/spdk_repo/spdk/test/bdev /home/vagrant/spdk_repo/spdk 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@335 -- # trap 'rm -f ./*.state; popd; exit 1' SIGINT SIGTERM EXIT 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # echo '' 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # sed s/--env-context=// 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # env_context= 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@339 -- # fio_config_gen /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio verify AIO '' 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1284 -- # local config_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1285 -- # local workload=verify 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1286 -- # local bdev_type=AIO 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1287 -- # local env_context= 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1288 -- # local fio_dir=/usr/src/fio 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1290 -- # '[' -e /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio ']' 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1295 -- # '[' -z verify ']' 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1299 -- # '[' -n '' ']' 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1303 -- # touch /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1305 -- # cat 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1317 -- # '[' verify == verify ']' 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1318 -- # cat 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1327 -- # '[' AIO == AIO ']' 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1328 -- # /usr/src/fio/fio --version 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1328 -- # [[ fio-3.35 == *\f\i\o\-\3* ]] 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1329 -- # echo serialize_overlap=1 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@340 -- # for b in "${bdevs_name[@]}" 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@341 -- # echo '[job_raid5f]' 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@342 -- # echo filename=raid5f 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@346 -- # local 'fio_params=--ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json' 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@348 -- # run_test bdev_fio_rw_verify fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1105 -- # '[' 11 -le 1 ']' 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:17:21.426 ************************************ 00:17:21.426 START TEST bdev_fio_rw_verify 00:17:21.426 ************************************ 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1129 -- # fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1360 -- # fio_plugin /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1341 -- # local fio_dir=/usr/src/fio 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1343 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1343 -- # local sanitizers 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1344 -- # local plugin=/home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # shift 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1347 -- # local asan_lib= 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1349 -- # ldd /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1349 -- # grep libasan 00:17:21.426 15:21:49 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:17:21.686 15:21:49 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1349 -- # asan_lib=/usr/lib64/libasan.so.8 00:17:21.686 15:21:49 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1350 -- # [[ -n /usr/lib64/libasan.so.8 ]] 00:17:21.686 15:21:49 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1351 -- # break 00:17:21.686 15:21:49 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1356 -- # LD_PRELOAD='/usr/lib64/libasan.so.8 /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev' 00:17:21.686 15:21:49 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1356 -- # /usr/src/fio/fio --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:17:21.686 job_raid5f: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:17:21.686 fio-3.35 00:17:21.686 Starting 1 thread 00:17:33.961 00:17:33.961 job_raid5f: (groupid=0, jobs=1): err= 0: pid=100776: Wed Nov 27 15:22:00 2024 00:17:33.961 read: IOPS=12.3k, BW=48.2MiB/s (50.5MB/s)(482MiB/10001msec) 00:17:33.961 slat (usec): min=17, max=106, avg=19.45, stdev= 1.92 00:17:33.961 clat (usec): min=11, max=316, avg=131.41, stdev=46.04 00:17:33.961 lat (usec): min=31, max=344, avg=150.86, stdev=46.31 00:17:33.961 clat percentiles (usec): 00:17:33.961 | 50.000th=[ 135], 99.000th=[ 219], 99.900th=[ 247], 99.990th=[ 277], 00:17:33.961 | 99.999th=[ 314] 00:17:33.961 write: IOPS=12.9k, BW=50.5MiB/s (52.9MB/s)(498MiB/9874msec); 0 zone resets 00:17:33.961 slat (usec): min=7, max=257, avg=16.14, stdev= 3.65 00:17:33.961 clat (usec): min=57, max=1793, avg=298.10, stdev=44.01 00:17:33.961 lat (usec): min=72, max=2050, avg=314.24, stdev=45.32 00:17:33.961 clat percentiles (usec): 00:17:33.961 | 50.000th=[ 302], 99.000th=[ 383], 99.900th=[ 635], 99.990th=[ 1450], 00:17:33.961 | 99.999th=[ 1729] 00:17:33.961 bw ( KiB/s): min=47248, max=55024, per=98.75%, avg=51021.89, stdev=1808.83, samples=19 00:17:33.961 iops : min=11812, max=13756, avg=12755.47, stdev=452.21, samples=19 00:17:33.961 lat (usec) : 20=0.01%, 50=0.01%, 100=15.87%, 250=39.20%, 500=44.84% 00:17:33.961 lat (usec) : 750=0.05%, 1000=0.02% 00:17:33.961 lat (msec) : 2=0.02% 00:17:33.961 cpu : usr=98.81%, sys=0.53%, ctx=34, majf=0, minf=13179 00:17:33.961 IO depths : 1=7.6%, 2=19.7%, 4=55.3%, 8=17.4%, 16=0.0%, 32=0.0%, >=64=0.0% 00:17:33.961 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:33.961 complete : 0=0.0%, 4=90.0%, 8=10.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:33.961 issued rwts: total=123403,127534,0,0 short=0,0,0,0 dropped=0,0,0,0 00:17:33.961 latency : target=0, window=0, percentile=100.00%, depth=8 00:17:33.961 00:17:33.961 Run status group 0 (all jobs): 00:17:33.961 READ: bw=48.2MiB/s (50.5MB/s), 48.2MiB/s-48.2MiB/s (50.5MB/s-50.5MB/s), io=482MiB (505MB), run=10001-10001msec 00:17:33.961 WRITE: bw=50.5MiB/s (52.9MB/s), 50.5MiB/s-50.5MiB/s (52.9MB/s-52.9MB/s), io=498MiB (522MB), run=9874-9874msec 00:17:33.961 ----------------------------------------------------- 00:17:33.961 Suppressions used: 00:17:33.961 count bytes template 00:17:33.961 1 7 /usr/src/fio/parse.c 00:17:33.961 312 29952 /usr/src/fio/iolog.c 00:17:33.961 1 8 libtcmalloc_minimal.so 00:17:33.961 1 904 libcrypto.so 00:17:33.961 ----------------------------------------------------- 00:17:33.961 00:17:33.961 00:17:33.961 real 0m11.427s 00:17:33.961 user 0m11.628s 00:17:33.961 sys 0m0.689s 00:17:33.961 15:22:00 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:33.961 15:22:00 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@10 -- # set +x 00:17:33.961 ************************************ 00:17:33.961 END TEST bdev_fio_rw_verify 00:17:33.961 ************************************ 00:17:33.961 15:22:00 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@349 -- # rm -f 00:17:33.961 15:22:00 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@350 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:17:33.961 15:22:00 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@353 -- # fio_config_gen /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio trim '' '' 00:17:33.961 15:22:00 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1284 -- # local config_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:17:33.961 15:22:00 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1285 -- # local workload=trim 00:17:33.961 15:22:00 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1286 -- # local bdev_type= 00:17:33.961 15:22:00 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1287 -- # local env_context= 00:17:33.961 15:22:00 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1288 -- # local fio_dir=/usr/src/fio 00:17:33.961 15:22:00 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1290 -- # '[' -e /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio ']' 00:17:33.961 15:22:00 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1295 -- # '[' -z trim ']' 00:17:33.961 15:22:00 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1299 -- # '[' -n '' ']' 00:17:33.961 15:22:00 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1303 -- # touch /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:17:33.961 15:22:01 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1305 -- # cat 00:17:33.961 15:22:01 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1317 -- # '[' trim == verify ']' 00:17:33.961 15:22:01 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1332 -- # '[' trim == trim ']' 00:17:33.961 15:22:01 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1333 -- # echo rw=trimwrite 00:17:33.961 15:22:01 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # jq -r 'select(.supported_io_types.unmap == true) | .name' 00:17:33.961 15:22:01 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # printf '%s\n' '{' ' "name": "raid5f",' ' "aliases": [' ' "f18e5124-227c-4909-b5ec-1ecded510ce2"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 131072,' ' "uuid": "f18e5124-227c-4909-b5ec-1ecded510ce2",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": false,' ' "flush": false,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "raid": {' ' "uuid": "f18e5124-227c-4909-b5ec-1ecded510ce2",' ' "strip_size_kb": 2,' ' "state": "online",' ' "raid_level": "raid5f",' ' "superblock": false,' ' "num_base_bdevs": 3,' ' "num_base_bdevs_discovered": 3,' ' "num_base_bdevs_operational": 3,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc0",' ' "uuid": "514fae21-1a69-422a-be2a-da58c5d8630a",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc1",' ' "uuid": "0bfe0434-02fc-4d29-a853-8a8573454675",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc2",' ' "uuid": "eb56b52a-0b23-456a-b81a-8f03b15a6b50",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' 00:17:33.961 15:22:01 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # [[ -n '' ]] 00:17:33.961 15:22:01 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@360 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:17:33.961 /home/vagrant/spdk_repo/spdk 00:17:33.961 ************************************ 00:17:33.961 END TEST bdev_fio 00:17:33.961 ************************************ 00:17:33.961 15:22:01 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@361 -- # popd 00:17:33.961 15:22:01 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@362 -- # trap - SIGINT SIGTERM EXIT 00:17:33.961 15:22:01 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@363 -- # return 0 00:17:33.961 00:17:33.961 real 0m11.698s 00:17:33.961 user 0m11.738s 00:17:33.961 sys 0m0.826s 00:17:33.961 15:22:01 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:33.961 15:22:01 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:17:33.961 15:22:01 blockdev_raid5f -- bdev/blockdev.sh@812 -- # trap cleanup SIGINT SIGTERM EXIT 00:17:33.961 15:22:01 blockdev_raid5f -- bdev/blockdev.sh@814 -- # run_test bdev_verify /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w verify -t 5 -C -m 0x3 '' 00:17:33.961 15:22:01 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 16 -le 1 ']' 00:17:33.961 15:22:01 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:33.961 15:22:01 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:33.961 ************************************ 00:17:33.961 START TEST bdev_verify 00:17:33.962 ************************************ 00:17:33.962 15:22:01 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w verify -t 5 -C -m 0x3 '' 00:17:33.962 [2024-11-27 15:22:01.226559] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:17:33.962 [2024-11-27 15:22:01.226713] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100928 ] 00:17:33.962 [2024-11-27 15:22:01.406776] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:17:33.962 [2024-11-27 15:22:01.451888] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:33.962 [2024-11-27 15:22:01.451967] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:17:33.962 Running I/O for 5 seconds... 00:17:35.842 10723.00 IOPS, 41.89 MiB/s [2024-11-27T15:22:04.890Z] 10789.50 IOPS, 42.15 MiB/s [2024-11-27T15:22:05.831Z] 10809.00 IOPS, 42.22 MiB/s [2024-11-27T15:22:06.770Z] 10873.25 IOPS, 42.47 MiB/s [2024-11-27T15:22:06.770Z] 10890.80 IOPS, 42.54 MiB/s 00:17:38.663 Latency(us) 00:17:38.663 [2024-11-27T15:22:06.770Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:17:38.663 Job: raid5f (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:17:38.663 Verification LBA range: start 0x0 length 0x2000 00:17:38.663 raid5f : 5.02 6601.43 25.79 0.00 0.00 29171.47 352.36 20834.15 00:17:38.663 Job: raid5f (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:17:38.663 Verification LBA range: start 0x2000 length 0x2000 00:17:38.663 raid5f : 5.02 4297.14 16.79 0.00 0.00 44789.26 129.68 32281.49 00:17:38.663 [2024-11-27T15:22:06.770Z] =================================================================================================================== 00:17:38.663 [2024-11-27T15:22:06.770Z] Total : 10898.57 42.57 0.00 0.00 35334.09 129.68 32281.49 00:17:39.234 00:17:39.234 real 0m5.972s 00:17:39.234 user 0m11.002s 00:17:39.234 sys 0m0.341s 00:17:39.234 15:22:07 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:39.234 15:22:07 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@10 -- # set +x 00:17:39.234 ************************************ 00:17:39.234 END TEST bdev_verify 00:17:39.234 ************************************ 00:17:39.234 15:22:07 blockdev_raid5f -- bdev/blockdev.sh@815 -- # run_test bdev_verify_big_io /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 65536 -w verify -t 5 -C -m 0x3 '' 00:17:39.234 15:22:07 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 16 -le 1 ']' 00:17:39.234 15:22:07 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:39.234 15:22:07 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:39.234 ************************************ 00:17:39.234 START TEST bdev_verify_big_io 00:17:39.234 ************************************ 00:17:39.234 15:22:07 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 65536 -w verify -t 5 -C -m 0x3 '' 00:17:39.234 [2024-11-27 15:22:07.270263] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:17:39.234 [2024-11-27 15:22:07.270380] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid101010 ] 00:17:39.498 [2024-11-27 15:22:07.440722] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:17:39.498 [2024-11-27 15:22:07.488012] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:39.498 [2024-11-27 15:22:07.488113] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:17:39.762 Running I/O for 5 seconds... 00:17:41.710 633.00 IOPS, 39.56 MiB/s [2024-11-27T15:22:11.196Z] 761.00 IOPS, 47.56 MiB/s [2024-11-27T15:22:12.137Z] 782.00 IOPS, 48.88 MiB/s [2024-11-27T15:22:13.076Z] 792.75 IOPS, 49.55 MiB/s [2024-11-27T15:22:13.076Z] 787.00 IOPS, 49.19 MiB/s 00:17:44.969 Latency(us) 00:17:44.969 [2024-11-27T15:22:13.076Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:17:44.969 Job: raid5f (Core Mask 0x1, workload: verify, depth: 128, IO size: 65536) 00:17:44.969 Verification LBA range: start 0x0 length 0x200 00:17:44.969 raid5f : 5.20 464.37 29.02 0.00 0.00 6921357.65 187.81 302209.68 00:17:44.969 Job: raid5f (Core Mask 0x2, workload: verify, depth: 128, IO size: 65536) 00:17:44.969 Verification LBA range: start 0x200 length 0x200 00:17:44.969 raid5f : 5.26 349.99 21.87 0.00 0.00 9002157.67 181.55 397451.51 00:17:44.969 [2024-11-27T15:22:13.076Z] =================================================================================================================== 00:17:44.969 [2024-11-27T15:22:13.076Z] Total : 814.35 50.90 0.00 0.00 7821586.21 181.55 397451.51 00:17:45.539 00:17:45.539 real 0m6.190s 00:17:45.539 user 0m11.465s 00:17:45.539 sys 0m0.323s 00:17:45.539 15:22:13 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:45.539 15:22:13 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@10 -- # set +x 00:17:45.539 ************************************ 00:17:45.539 END TEST bdev_verify_big_io 00:17:45.539 ************************************ 00:17:45.539 15:22:13 blockdev_raid5f -- bdev/blockdev.sh@816 -- # run_test bdev_write_zeroes /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:45.539 15:22:13 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 13 -le 1 ']' 00:17:45.540 15:22:13 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:45.540 15:22:13 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:45.540 ************************************ 00:17:45.540 START TEST bdev_write_zeroes 00:17:45.540 ************************************ 00:17:45.540 15:22:13 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:45.540 [2024-11-27 15:22:13.525601] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:17:45.540 [2024-11-27 15:22:13.525721] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid101097 ] 00:17:45.800 [2024-11-27 15:22:13.696020] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:45.800 [2024-11-27 15:22:13.741930] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:46.060 Running I/O for 1 seconds... 00:17:47.000 29679.00 IOPS, 115.93 MiB/s 00:17:47.000 Latency(us) 00:17:47.000 [2024-11-27T15:22:15.107Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:17:47.000 Job: raid5f (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:17:47.000 raid5f : 1.01 29644.31 115.80 0.00 0.00 4304.21 1459.54 5866.76 00:17:47.000 [2024-11-27T15:22:15.107Z] =================================================================================================================== 00:17:47.000 [2024-11-27T15:22:15.107Z] Total : 29644.31 115.80 0.00 0.00 4304.21 1459.54 5866.76 00:17:47.260 00:17:47.260 real 0m1.921s 00:17:47.260 user 0m1.506s 00:17:47.260 sys 0m0.304s 00:17:47.260 15:22:15 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:47.260 15:22:15 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@10 -- # set +x 00:17:47.260 ************************************ 00:17:47.260 END TEST bdev_write_zeroes 00:17:47.260 ************************************ 00:17:47.520 15:22:15 blockdev_raid5f -- bdev/blockdev.sh@819 -- # run_test bdev_json_nonenclosed /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:47.520 15:22:15 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 13 -le 1 ']' 00:17:47.520 15:22:15 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:47.520 15:22:15 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:47.520 ************************************ 00:17:47.520 START TEST bdev_json_nonenclosed 00:17:47.520 ************************************ 00:17:47.520 15:22:15 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:47.520 [2024-11-27 15:22:15.528275] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:17:47.520 [2024-11-27 15:22:15.528409] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid101134 ] 00:17:47.780 [2024-11-27 15:22:15.702039] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:47.780 [2024-11-27 15:22:15.745875] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:47.780 [2024-11-27 15:22:15.746007] json_config.c: 608:json_config_prepare_ctx: *ERROR*: Invalid JSON configuration: not enclosed in {}. 00:17:47.780 [2024-11-27 15:22:15.746031] rpc.c: 190:spdk_rpc_server_finish: *ERROR*: No server listening on provided address: 00:17:47.780 [2024-11-27 15:22:15.746045] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:17:47.780 00:17:47.780 real 0m0.415s 00:17:47.780 user 0m0.171s 00:17:47.780 sys 0m0.140s 00:17:47.780 15:22:15 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:47.780 15:22:15 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@10 -- # set +x 00:17:47.780 ************************************ 00:17:47.780 END TEST bdev_json_nonenclosed 00:17:47.780 ************************************ 00:17:48.040 15:22:15 blockdev_raid5f -- bdev/blockdev.sh@822 -- # run_test bdev_json_nonarray /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:48.040 15:22:15 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 13 -le 1 ']' 00:17:48.040 15:22:15 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:48.040 15:22:15 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:48.040 ************************************ 00:17:48.040 START TEST bdev_json_nonarray 00:17:48.040 ************************************ 00:17:48.040 15:22:15 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:48.041 [2024-11-27 15:22:16.013510] Starting SPDK v25.01-pre git sha1 35cd3e84d / DPDK 23.11.0 initialization... 00:17:48.041 [2024-11-27 15:22:16.013651] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid101160 ] 00:17:48.300 [2024-11-27 15:22:16.184301] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:48.300 [2024-11-27 15:22:16.228434] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:48.300 [2024-11-27 15:22:16.228579] json_config.c: 614:json_config_prepare_ctx: *ERROR*: Invalid JSON configuration: 'subsystems' should be an array. 00:17:48.300 [2024-11-27 15:22:16.228608] rpc.c: 190:spdk_rpc_server_finish: *ERROR*: No server listening on provided address: 00:17:48.300 [2024-11-27 15:22:16.228622] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:17:48.300 00:17:48.300 real 0m0.410s 00:17:48.300 user 0m0.169s 00:17:48.300 sys 0m0.137s 00:17:48.300 15:22:16 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:48.300 15:22:16 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@10 -- # set +x 00:17:48.300 ************************************ 00:17:48.300 END TEST bdev_json_nonarray 00:17:48.300 ************************************ 00:17:48.300 15:22:16 blockdev_raid5f -- bdev/blockdev.sh@824 -- # [[ raid5f == bdev ]] 00:17:48.300 15:22:16 blockdev_raid5f -- bdev/blockdev.sh@832 -- # [[ raid5f == gpt ]] 00:17:48.300 15:22:16 blockdev_raid5f -- bdev/blockdev.sh@836 -- # [[ raid5f == crypto_sw ]] 00:17:48.300 15:22:16 blockdev_raid5f -- bdev/blockdev.sh@848 -- # trap - SIGINT SIGTERM EXIT 00:17:48.300 15:22:16 blockdev_raid5f -- bdev/blockdev.sh@849 -- # cleanup 00:17:48.301 15:22:16 blockdev_raid5f -- bdev/blockdev.sh@23 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/aiofile 00:17:48.301 15:22:16 blockdev_raid5f -- bdev/blockdev.sh@24 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:17:48.560 15:22:16 blockdev_raid5f -- bdev/blockdev.sh@26 -- # [[ raid5f == rbd ]] 00:17:48.560 15:22:16 blockdev_raid5f -- bdev/blockdev.sh@30 -- # [[ raid5f == daos ]] 00:17:48.560 15:22:16 blockdev_raid5f -- bdev/blockdev.sh@34 -- # [[ raid5f = \g\p\t ]] 00:17:48.560 15:22:16 blockdev_raid5f -- bdev/blockdev.sh@40 -- # [[ raid5f == xnvme ]] 00:17:48.560 00:17:48.560 real 0m36.390s 00:17:48.560 user 0m48.731s 00:17:48.560 sys 0m5.387s 00:17:48.560 15:22:16 blockdev_raid5f -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:48.560 15:22:16 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:48.560 ************************************ 00:17:48.560 END TEST blockdev_raid5f 00:17:48.560 ************************************ 00:17:48.560 15:22:16 -- spdk/autotest.sh@194 -- # uname -s 00:17:48.560 15:22:16 -- spdk/autotest.sh@194 -- # [[ Linux == Linux ]] 00:17:48.560 15:22:16 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:17:48.560 15:22:16 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:17:48.560 15:22:16 -- spdk/autotest.sh@207 -- # '[' 0 -eq 1 ']' 00:17:48.560 15:22:16 -- spdk/autotest.sh@256 -- # '[' 0 -eq 1 ']' 00:17:48.560 15:22:16 -- spdk/autotest.sh@260 -- # timing_exit lib 00:17:48.560 15:22:16 -- common/autotest_common.sh@732 -- # xtrace_disable 00:17:48.560 15:22:16 -- common/autotest_common.sh@10 -- # set +x 00:17:48.560 15:22:16 -- spdk/autotest.sh@262 -- # '[' 0 -eq 1 ']' 00:17:48.560 15:22:16 -- spdk/autotest.sh@267 -- # '[' 0 -eq 1 ']' 00:17:48.560 15:22:16 -- spdk/autotest.sh@276 -- # '[' 0 -eq 1 ']' 00:17:48.560 15:22:16 -- spdk/autotest.sh@311 -- # '[' 0 -eq 1 ']' 00:17:48.560 15:22:16 -- spdk/autotest.sh@315 -- # '[' 0 -eq 1 ']' 00:17:48.560 15:22:16 -- spdk/autotest.sh@319 -- # '[' 0 -eq 1 ']' 00:17:48.560 15:22:16 -- spdk/autotest.sh@324 -- # '[' 0 -eq 1 ']' 00:17:48.560 15:22:16 -- spdk/autotest.sh@333 -- # '[' 0 -eq 1 ']' 00:17:48.560 15:22:16 -- spdk/autotest.sh@338 -- # '[' 0 -eq 1 ']' 00:17:48.560 15:22:16 -- spdk/autotest.sh@342 -- # '[' 0 -eq 1 ']' 00:17:48.560 15:22:16 -- spdk/autotest.sh@346 -- # '[' 0 -eq 1 ']' 00:17:48.560 15:22:16 -- spdk/autotest.sh@350 -- # '[' 0 -eq 1 ']' 00:17:48.560 15:22:16 -- spdk/autotest.sh@355 -- # '[' 0 -eq 1 ']' 00:17:48.560 15:22:16 -- spdk/autotest.sh@359 -- # '[' 0 -eq 1 ']' 00:17:48.560 15:22:16 -- spdk/autotest.sh@366 -- # [[ 0 -eq 1 ]] 00:17:48.560 15:22:16 -- spdk/autotest.sh@370 -- # [[ 0 -eq 1 ]] 00:17:48.560 15:22:16 -- spdk/autotest.sh@374 -- # [[ 0 -eq 1 ]] 00:17:48.560 15:22:16 -- spdk/autotest.sh@378 -- # [[ '' -eq 1 ]] 00:17:48.560 15:22:16 -- spdk/autotest.sh@385 -- # trap - SIGINT SIGTERM EXIT 00:17:48.560 15:22:16 -- spdk/autotest.sh@387 -- # timing_enter post_cleanup 00:17:48.560 15:22:16 -- common/autotest_common.sh@726 -- # xtrace_disable 00:17:48.560 15:22:16 -- common/autotest_common.sh@10 -- # set +x 00:17:48.560 15:22:16 -- spdk/autotest.sh@388 -- # autotest_cleanup 00:17:48.560 15:22:16 -- common/autotest_common.sh@1396 -- # local autotest_es=0 00:17:48.560 15:22:16 -- common/autotest_common.sh@1397 -- # xtrace_disable 00:17:48.560 15:22:16 -- common/autotest_common.sh@10 -- # set +x 00:17:51.101 INFO: APP EXITING 00:17:51.101 INFO: killing all VMs 00:17:51.101 INFO: killing vhost app 00:17:51.101 INFO: EXIT DONE 00:17:51.363 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:17:51.363 Waiting for block devices as requested 00:17:51.622 0000:00:11.0 (1b36 0010): uio_pci_generic -> nvme 00:17:51.622 0000:00:10.0 (1b36 0010): uio_pci_generic -> nvme 00:17:52.564 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:17:52.564 Cleaning 00:17:52.564 Removing: /var/run/dpdk/spdk0/config 00:17:52.564 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-0 00:17:52.564 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-1 00:17:52.564 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-2 00:17:52.564 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-3 00:17:52.564 Removing: /var/run/dpdk/spdk0/fbarray_memzone 00:17:52.564 Removing: /var/run/dpdk/spdk0/hugepage_info 00:17:52.564 Removing: /dev/shm/spdk_tgt_trace.pid69301 00:17:52.564 Removing: /var/run/dpdk/spdk0 00:17:52.564 Removing: /var/run/dpdk/spdk_pid100207 00:17:52.564 Removing: /var/run/dpdk/spdk_pid100465 00:17:52.564 Removing: /var/run/dpdk/spdk_pid100510 00:17:52.564 Removing: /var/run/dpdk/spdk_pid100541 00:17:52.564 Removing: /var/run/dpdk/spdk_pid100765 00:17:52.564 Removing: /var/run/dpdk/spdk_pid100928 00:17:52.564 Removing: /var/run/dpdk/spdk_pid101010 00:17:52.564 Removing: /var/run/dpdk/spdk_pid101097 00:17:52.564 Removing: /var/run/dpdk/spdk_pid101134 00:17:52.564 Removing: /var/run/dpdk/spdk_pid101160 00:17:52.564 Removing: /var/run/dpdk/spdk_pid69138 00:17:52.564 Removing: /var/run/dpdk/spdk_pid69301 00:17:52.564 Removing: /var/run/dpdk/spdk_pid69508 00:17:52.564 Removing: /var/run/dpdk/spdk_pid69596 00:17:52.564 Removing: /var/run/dpdk/spdk_pid69619 00:17:52.564 Removing: /var/run/dpdk/spdk_pid69736 00:17:52.564 Removing: /var/run/dpdk/spdk_pid69754 00:17:52.564 Removing: /var/run/dpdk/spdk_pid69933 00:17:52.564 Removing: /var/run/dpdk/spdk_pid70010 00:17:52.564 Removing: /var/run/dpdk/spdk_pid70095 00:17:52.564 Removing: /var/run/dpdk/spdk_pid70195 00:17:52.564 Removing: /var/run/dpdk/spdk_pid70270 00:17:52.824 Removing: /var/run/dpdk/spdk_pid70315 00:17:52.824 Removing: /var/run/dpdk/spdk_pid70346 00:17:52.824 Removing: /var/run/dpdk/spdk_pid70422 00:17:52.824 Removing: /var/run/dpdk/spdk_pid70517 00:17:52.824 Removing: /var/run/dpdk/spdk_pid70943 00:17:52.824 Removing: /var/run/dpdk/spdk_pid70992 00:17:52.825 Removing: /var/run/dpdk/spdk_pid71044 00:17:52.825 Removing: /var/run/dpdk/spdk_pid71060 00:17:52.825 Removing: /var/run/dpdk/spdk_pid71119 00:17:52.825 Removing: /var/run/dpdk/spdk_pid71135 00:17:52.825 Removing: /var/run/dpdk/spdk_pid71204 00:17:52.825 Removing: /var/run/dpdk/spdk_pid71214 00:17:52.825 Removing: /var/run/dpdk/spdk_pid71262 00:17:52.825 Removing: /var/run/dpdk/spdk_pid71280 00:17:52.825 Removing: /var/run/dpdk/spdk_pid71322 00:17:52.825 Removing: /var/run/dpdk/spdk_pid71340 00:17:52.825 Removing: /var/run/dpdk/spdk_pid71469 00:17:52.825 Removing: /var/run/dpdk/spdk_pid71511 00:17:52.825 Removing: /var/run/dpdk/spdk_pid71589 00:17:52.825 Removing: /var/run/dpdk/spdk_pid72757 00:17:52.825 Removing: /var/run/dpdk/spdk_pid72952 00:17:52.825 Removing: /var/run/dpdk/spdk_pid73081 00:17:52.825 Removing: /var/run/dpdk/spdk_pid73691 00:17:52.825 Removing: /var/run/dpdk/spdk_pid73886 00:17:52.825 Removing: /var/run/dpdk/spdk_pid74015 00:17:52.825 Removing: /var/run/dpdk/spdk_pid74620 00:17:52.825 Removing: /var/run/dpdk/spdk_pid74939 00:17:52.825 Removing: /var/run/dpdk/spdk_pid75068 00:17:52.825 Removing: /var/run/dpdk/spdk_pid76409 00:17:52.825 Removing: /var/run/dpdk/spdk_pid76651 00:17:52.825 Removing: /var/run/dpdk/spdk_pid76780 00:17:52.825 Removing: /var/run/dpdk/spdk_pid78121 00:17:52.825 Removing: /var/run/dpdk/spdk_pid78363 00:17:52.825 Removing: /var/run/dpdk/spdk_pid78492 00:17:52.825 Removing: /var/run/dpdk/spdk_pid79822 00:17:52.825 Removing: /var/run/dpdk/spdk_pid80257 00:17:52.825 Removing: /var/run/dpdk/spdk_pid80391 00:17:52.825 Removing: /var/run/dpdk/spdk_pid81816 00:17:52.825 Removing: /var/run/dpdk/spdk_pid82064 00:17:52.825 Removing: /var/run/dpdk/spdk_pid82194 00:17:52.825 Removing: /var/run/dpdk/spdk_pid83617 00:17:52.825 Removing: /var/run/dpdk/spdk_pid83861 00:17:52.825 Removing: /var/run/dpdk/spdk_pid83990 00:17:52.825 Removing: /var/run/dpdk/spdk_pid85420 00:17:52.825 Removing: /var/run/dpdk/spdk_pid85896 00:17:52.825 Removing: /var/run/dpdk/spdk_pid86030 00:17:52.825 Removing: /var/run/dpdk/spdk_pid86158 00:17:52.825 Removing: /var/run/dpdk/spdk_pid86572 00:17:52.825 Removing: /var/run/dpdk/spdk_pid87294 00:17:52.825 Removing: /var/run/dpdk/spdk_pid87680 00:17:52.825 Removing: /var/run/dpdk/spdk_pid88356 00:17:52.825 Removing: /var/run/dpdk/spdk_pid88793 00:17:52.825 Removing: /var/run/dpdk/spdk_pid89530 00:17:52.825 Removing: /var/run/dpdk/spdk_pid89929 00:17:52.825 Removing: /var/run/dpdk/spdk_pid91843 00:17:52.825 Removing: /var/run/dpdk/spdk_pid92270 00:17:53.085 Removing: /var/run/dpdk/spdk_pid92699 00:17:53.085 Removing: /var/run/dpdk/spdk_pid94736 00:17:53.085 Removing: /var/run/dpdk/spdk_pid95215 00:17:53.085 Removing: /var/run/dpdk/spdk_pid95720 00:17:53.085 Removing: /var/run/dpdk/spdk_pid96754 00:17:53.085 Removing: /var/run/dpdk/spdk_pid97071 00:17:53.085 Removing: /var/run/dpdk/spdk_pid97986 00:17:53.085 Removing: /var/run/dpdk/spdk_pid98302 00:17:53.085 Removing: /var/run/dpdk/spdk_pid99223 00:17:53.085 Removing: /var/run/dpdk/spdk_pid99540 00:17:53.085 Clean 00:17:53.085 15:22:21 -- common/autotest_common.sh@1453 -- # return 0 00:17:53.085 15:22:21 -- spdk/autotest.sh@389 -- # timing_exit post_cleanup 00:17:53.085 15:22:21 -- common/autotest_common.sh@732 -- # xtrace_disable 00:17:53.085 15:22:21 -- common/autotest_common.sh@10 -- # set +x 00:17:53.085 15:22:21 -- spdk/autotest.sh@391 -- # timing_exit autotest 00:17:53.085 15:22:21 -- common/autotest_common.sh@732 -- # xtrace_disable 00:17:53.085 15:22:21 -- common/autotest_common.sh@10 -- # set +x 00:17:53.085 15:22:21 -- spdk/autotest.sh@392 -- # chmod a+r /home/vagrant/spdk_repo/spdk/../output/timing.txt 00:17:53.085 15:22:21 -- spdk/autotest.sh@394 -- # [[ -f /home/vagrant/spdk_repo/spdk/../output/udev.log ]] 00:17:53.085 15:22:21 -- spdk/autotest.sh@394 -- # rm -f /home/vagrant/spdk_repo/spdk/../output/udev.log 00:17:53.085 15:22:21 -- spdk/autotest.sh@396 -- # [[ y == y ]] 00:17:53.085 15:22:21 -- spdk/autotest.sh@398 -- # hostname 00:17:53.085 15:22:21 -- spdk/autotest.sh@398 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -c --no-external -d /home/vagrant/spdk_repo/spdk -t fedora39-cloud-1721788873-2326 -o /home/vagrant/spdk_repo/spdk/../output/cov_test.info 00:17:53.346 geninfo: WARNING: invalid characters removed from testname! 00:18:15.339 15:22:42 -- spdk/autotest.sh@399 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -a /home/vagrant/spdk_repo/spdk/../output/cov_base.info -a /home/vagrant/spdk_repo/spdk/../output/cov_test.info -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:18:17.248 15:22:45 -- spdk/autotest.sh@400 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/dpdk/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:18:19.156 15:22:47 -- spdk/autotest.sh@404 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info --ignore-errors unused,unused '/usr/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:18:21.065 15:22:49 -- spdk/autotest.sh@405 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/examples/vmd/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:18:23.602 15:22:51 -- spdk/autotest.sh@406 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/app/spdk_lspci/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:18:24.984 15:22:53 -- spdk/autotest.sh@407 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/app/spdk_top/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:18:27.526 15:22:55 -- spdk/autotest.sh@408 -- # rm -f cov_base.info cov_test.info OLD_STDOUT OLD_STDERR 00:18:27.526 15:22:55 -- spdk/autorun.sh@1 -- $ timing_finish 00:18:27.526 15:22:55 -- common/autotest_common.sh@738 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/timing.txt ]] 00:18:27.526 15:22:55 -- common/autotest_common.sh@740 -- $ flamegraph=/usr/local/FlameGraph/flamegraph.pl 00:18:27.526 15:22:55 -- common/autotest_common.sh@741 -- $ [[ -x /usr/local/FlameGraph/flamegraph.pl ]] 00:18:27.526 15:22:55 -- common/autotest_common.sh@744 -- $ /usr/local/FlameGraph/flamegraph.pl --title 'Build Timing' --nametype Step: --countname seconds /home/vagrant/spdk_repo/spdk/../output/timing.txt 00:18:27.526 + [[ -n 6172 ]] 00:18:27.526 + sudo kill 6172 00:18:27.536 [Pipeline] } 00:18:27.553 [Pipeline] // timeout 00:18:27.561 [Pipeline] } 00:18:27.578 [Pipeline] // stage 00:18:27.584 [Pipeline] } 00:18:27.600 [Pipeline] // catchError 00:18:27.611 [Pipeline] stage 00:18:27.614 [Pipeline] { (Stop VM) 00:18:27.629 [Pipeline] sh 00:18:27.918 + vagrant halt 00:18:30.460 ==> default: Halting domain... 00:18:38.608 [Pipeline] sh 00:18:38.893 + vagrant destroy -f 00:18:41.434 ==> default: Removing domain... 00:18:41.446 [Pipeline] sh 00:18:41.729 + mv output /var/jenkins/workspace/raid-vg-autotest/output 00:18:41.739 [Pipeline] } 00:18:41.758 [Pipeline] // stage 00:18:41.764 [Pipeline] } 00:18:41.781 [Pipeline] // dir 00:18:41.786 [Pipeline] } 00:18:41.803 [Pipeline] // wrap 00:18:41.809 [Pipeline] } 00:18:41.823 [Pipeline] // catchError 00:18:41.832 [Pipeline] stage 00:18:41.835 [Pipeline] { (Epilogue) 00:18:41.849 [Pipeline] sh 00:18:42.138 + jbp/jenkins/jjb-config/jobs/scripts/compress_artifacts.sh 00:18:46.395 [Pipeline] catchError 00:18:46.398 [Pipeline] { 00:18:46.413 [Pipeline] sh 00:18:46.700 + jbp/jenkins/jjb-config/jobs/scripts/check_artifacts_size.sh 00:18:46.700 Artifacts sizes are good 00:18:46.711 [Pipeline] } 00:18:46.726 [Pipeline] // catchError 00:18:46.740 [Pipeline] archiveArtifacts 00:18:46.748 Archiving artifacts 00:18:46.855 [Pipeline] cleanWs 00:18:46.871 [WS-CLEANUP] Deleting project workspace... 00:18:46.871 [WS-CLEANUP] Deferred wipeout is used... 00:18:46.878 [WS-CLEANUP] done 00:18:46.880 [Pipeline] } 00:18:46.898 [Pipeline] // stage 00:18:46.904 [Pipeline] } 00:18:46.918 [Pipeline] // node 00:18:46.923 [Pipeline] End of Pipeline 00:18:46.966 Finished: SUCCESS